ADJUSTMENT COMPUTATIONS
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
ADJUSTMENT
COMPUTATIONS
Spatial Data Analysis
Fourth Edition
CHARLES D. GHILANI, Ph.D.
Professor of Engineering
Surveying Engineering Program
Pennsylvania State University
PAUL R. WOLF, Ph.D.
Professor Emeritus
Department of Civil and Environmental Engineering
University of Wisconsin–Madison
JOHN WILEY & SONS, INC.
This book is printed on acid-free paper.
⬁
Copyright 2006 by John Wiley & Sons, Inc. All rights reserved.
Published by John Wiley & Sons, Inc., Hoboken, New Jersey
Published simultaneously in Canada.
No part of this publication may be reproduced, stored in a retrieval system, or transmitted in
any form or by any means, electronic, mechanical, photocopying, recording, scanning, or
otherwise, except as permitted under Section 107 or 108 of the 1976 United States Copyright
Act, without either the prior written permission of the Publisher, or authorization through
payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc., 222
Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 750-4470, or on the web at
www.copyright.com. Requests to the Publisher for permission should be addressed to the
Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201)
748-6011, fax (201) 748-6008, or online at http: / / www.wiley.com / go / permission.
Limit of Liability / Disclaimer of Warranty: While the publisher and author have used their best
efforts in preparing this book, they make no representations or warranties with respect to the
accuracy or completeness of the contents of this book and specifically disclaim any implied
warranties of merchantability or fitness for a particular purpose. No warranty may be created or
extended by sales representatives or written sales materials. The advice and strategies contained
herein may not be suitable for your situation. You should consult with a professional where
appropriate. Neither the publisher nor author shall be liable for any loss of profit or any other
commercial damages, including but not limited to special, incidental, consequential, or other
damages.
For general information on our other products and services or for technical support, please
contact our Customer Care Department within the United States at (800) 762-2974, outside the
United States at (317) 572-3993 or fax (317) 572-4002.
Wiley also publishes its books in a variety of electronic formats. Some content that appears in
print may not be available in electronic books. For more information about Wiley products,
visit our web site at www.wiley.com.
Library of Congress Cataloging-in-Publication Data:
Ghilani, Charles D.
Adjustment computations : spatial data analysis / Charles D. Ghilani, Paul
R. Wolf.—4th ed.
p. cm.
Prev. ed. entered under Wolf.
ISBN-13 978-0-471-69728-2 (cloth)
ISBN-10 0-471-69728-1 (cloth)
1. Surveying—Mathematics. 2. Spatial analysis (Statistics) I. Wolf, Paul
R. II. Title.
TA556.M38W65 2006
526.9—dc22
2005028948
Printed in the United States of America
10 9 8 7 6 5 4 3 2 1
CONTENTS
PREFACE
xix
ACKNOWLEDGMENTS
1
INTRODUCTION
xxiii
1
1.1
1.2
1.3
1.4
1.5
1.6
Introduction / 1
Direct and Indirect Measurements / 2
Measurement Error Sources / 2
Definitions / 3
Precision versus Accuracy / 4
Redundant Measurements in Surveying and Their
Adjustment / 7
1.7
Advantages of Least Squares Adjustment / 8
1.8
Overview of the Book / 10
Problems / 10
2
OBSERVATIONS AND THEIR ANALYSIS
2.1
2.2
2.3
2.4
2.5
2.6
2.7
12
Introduction / 12
Sample versus Population / 12
Range and Median / 13
Graphical Representation of Data / 14
Numerical Methods of Describing Data / 17
Measures of Central Tendency / 17
Additional Definitions / 18
v
vi
CONTENTS
2.8
2.9
2.10
Alternative Formula for Determining Variance / 21
Numerical Examples / 23
Derivation of the Sample Variance (Bessel’s
Correction) / 28
2.11 Programming / 29
Problems / 30
3
RANDOM ERROR THEORY
33
3.1
3.2
3.3
3.4
3.5
Introduction / 33
Theory of Probability / 33
Properties of the Normal Distribution Curve / 36
Standard Normal Distribution Function / 38
Probability of the Standard Error / 41
3.5.1 50% Probable Error / 42
3.5.2 95% Probable Error / 42
3.5.3 Other Percent Probable Errors / 43
3.6
Uses for Percent Errors / 43
3.7
Practical Examples / 44
Problems / 47
4
CONFIDENCE INTERVALS
50
4.1
4.2
Introduction / 50
Distributions Used in Sampling Theory / 52
4.2.1 2 Distribution / 52
4.2.2 t (Student) Distribution / 54
4.2.3 F Distribution / 55
4.3
Confidence Interval for the Mean: t Statistic / 56
4.4
Testing the Validity of the Confidence Interval / 59
4.5
Selecting a Sample Size / 60
4.6
Confidence Interval for a Population Variance / 61
4.7
Confidence Interval for the Ratio of Two Population
Variances / 63
Problems / 65
5
STATISTICAL TESTING
5.1
5.2
5.3
5.4
Hypothesis Testing / 68
Systematic Development of a Test / 71
Test of Hypothesis for the Population Mean / 72
Test of Hypothesis for the Population Variance / 74
68
CONTENTS
vii
5.5
Test of Hypothesis for the Ratio of Two Population
Variances / 77
Problems / 81
6
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY
MEASURED QUANTITIES
84
6.1
Basic Error Propagation Equation / 84
6.1.1 Generic Example / 88
6.2
Frequently Encountered Specific Functions / 88
6.2.1 Standard Deviation of a Sum / 88
6.2.2 Standard Deviation in a Series / 89
6.2.3 Standard Deviation of the Mean / 89
6.3
Numerical Examples / 89
6.4
Conclusions / 94
Problems / 95
7
ERROR PROPAGATION IN ANGLE AND DISTANCE
OBSERVATIONS
7.1
7.2
7.3
Introduction / 99
Error Sources in Horizontal Angles / 99
Reading Errors / 100
7.3.1 Angles Observed by the Repetition
Method / 100
7.3.2 Angles Observed by the Directional
Method / 101
7.4
Pointing Errors / 102
7.5
Estimated Pointing and Reading Errors with Total
Stations / 103
7.6
Target Centering Errors / 104
7.7
Instrument Centering Errors / 106
7.8
Effects of Leveling Errors in Angle Observations / 110
7.9
Numerical Example of Combined Error Propagation in
a Single Horizontal Angle / 112
7.10 Use of Estimated Errors to Check Angular Misclosure
in a Traverse / 114
7.11 Errors in Astronomical Observations for an
Azimuth / 116
7.12 Errors in Electronic Distance Observations / 121
7.13 Use of Computational Software / 123
Problems / 123
99
viii
8
CONTENTS
ERROR PROPAGATION IN TRAVERSE SURVEYS
127
8.1
8.2
Introduction / 127
Derivation of Estimated Error in Latitude and
Departure / 128
8.3
Derivation of Estimated Standard Errors in Course
Azimuths / 129
8.4
Computing and Analyzing Polygon Traverse Misclosure
Errors / 130
8.5
Computing and Analyzing Link Traverse Misclosure
Errors / 135
8.6
Conclusions / 140
Problems / 140
9
ERROR PROPAGATION IN ELEVATION DETERMINATION
144
9.1
9.2
Introduction / 144
Systematic Errors in Differential Leveling / 144
9.2.1 Collimation Error / 144
9.2.2 Earth Curvature and Refraction / 146
9.2.3 Combined Effects of Systematic Errors on
Elevation Differences / 147
9.3
Random Errors in Differential Leveling / 148
9.3.1 Reading Errors / 148
9.3.2 Instrument Leveling Errors / 148
9.3.3 Rod Plumbing Error / 148
9.3.4 Estimated Errors in Differential
Leveling / 150
9.4
Error Propagation in Trigonometric Leveling / 152
Problems / 156
10
WEIGHTS OF OBSERVATIONS
10.1
10.2
10.3
10.4
10.5
10.6
Introduction / 159
Weighted Mean / 161
Relation between Weights and Standard Errors / 163
Statistics of Weighted Observations / 164
10.4.1 Standard Deviation / 164
10.4.2 Standard Error of Weight w and Standard
Error of the Weighted Mean / 164
Weights in Angle Observations / 165
Weights in Differential Leveling / 166
159
CONTENTS
ix
10.7
Practical Examples / 167
Problems / 170
11
PRINCIPLES OF LEAST SQUARES
173
11.1
11.2
11.3
11.4
11.5
11.6
Introduction / 173
Fundamental Principle of Least Squares / 174
Fundamental Principle of Weighted Least Squares / 176
Stochastic Model / 177
Functional Model / 177
Observation Equations / 179
11.6.1 Elementary Example of Observation Equation
Adjustment / 179
11.7
Systematic Formulation of the Normal Equations / 181
11.7.1 Equal-Weight Case / 181
11.7.2 Weighted Case / 183
11.7.3 Advantages of the Systematic Approach / 184
11.8
Tabular Formation of the Normal Equations / 184
11.9
Using Matrices to Form the Normal Equations / 185
11.9.1 Equal-Weight Case / 185
11.9.2 Weighted Case / 187
11.10 Least Squares Solution of Nonlinear Systems / 188
11.11 Least Squares Fit of Points to a Line or Curve / 191
11.11.1 Fitting Data to a Straight Line / 192
11.11.2 Fitting Data to a Parabola / 194
11.12 Calibration of an EDM Instrument / 195
11.13 Least Squares Adjustment Using Conditional
Equations / 196
11.14 Example 11.5 Using Observation Equations / 198
Problems / 200
12
ADJUSTMENT OF LEVEL NETS
12.1
12.2
12.3
12.4
12.5
Introduction / 205
Observation Equation / 205
Unweighted Example / 206
Weighted Example / 209
Reference Standard Deviation / 211
12.5.1 Unweighted Example / 212
12.5.2 Weighted Example / 213
205
x
CONTENTS
12.6
Another Weighted Adjustment / 213
Problems / 216
13
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
221
13.1
Introduction / 221
13.2
Development of the Covariance Matrix / 221
13.3
Numerical Examples / 225
13.4
Standard Deviations of Computed Quantities / 226
Problems / 229
14
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION 233
14.1
14.2
14.3
14.4
Introduction / 233
Distance Observation Equation / 235
Trilateration Adjustment Example / 237
Formulation of a Generalized Coefficient Matrix for a
More Complex Network / 243
14.5
Computer Solution of a Trilaterated Quadrilateral / 244
14.6
Iteration Termination / 248
14.6.1 Method of Maximum Iterations / 249
14.6.2 Maximum Correction / 249
14.6.3 Monitoring the Adjustment’s Reference
Variance / 249
Problems / 250
15
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION 255
15.1
15.2
Introduction / 255
Azimuth Observation Equation / 255
15.2.1 Linearization of the Azimuth Observation
Equation / 256
15.3
Angle Observation Equation / 258
15.4
Adjustment of Intersections / 260
15.5
Adjustment of Resections / 265
15.5.1 Computing Initial Approximations in the
Resection Problem / 266
15.6
Adjustment of Triangulated Quadrilaterals / 271
Problems / 275
CONTENTS
16
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES
AND NETWORKS
xi
283
16.1
Introduction to Traverse Adjustments / 283
16.2
Observation Equations / 283
16.3
Redundant Equations / 284
16.4
Numerical Example / 285
16.5
Minimum Amount of Control / 291
16.6
Adjustment of Networks / 291
16.7 2 Test: Goodness of Fit / 300
Problems / 301
17
ADJUSTMENT OF GPS NETWORKS
310
17.1
17.2
17.3
17.4
Introduction / 310
GPS Observations / 311
GPS Errors and the Need for Adjustment / 314
Reference Coordinate Systems for GPS
Observations / 314
17.5
Converting between the Terrestrial and Geodetic
Coordinate Systems / 316
17.6
Application of Least Squares in Processing GPS
Data / 321
17.7
Network Preadjustment Data Analysis / 322
17.7.1 Analysis of Fixed Baseline Measurements / 322
17.7.2 Analysis of Repeat Baseline Measurements / 324
17.7.3 Analysis of Loop Closures / 325
17.7.4 Minimally Constrained Adjustment / 326
17.8
Least Squares Adjustment of GPS Networks / 327
Problems / 332
18
COORDINATE TRANSFORMATIONS
18.1
18.2
18.3
18.4
18.5
18.6
Introduction / 345
Two-Dimensional Conformal Coordinate
Transformation / 345
Equation Development / 346
Application of Least Squares / 348
Two-Dimensional Affine Coordinate Transformation / 350
Two-Dimensional Projective Coordinate
Transformation / 353
345
xii
CONTENTS
18.7
Three-Dimensional Conformal Coordinate
Transformation / 356
18.8
Statistically Valid Parameters / 362
Problems / 364
19
ERROR ELLIPSE
369
19.1
19.2
19.3
Introduction / 369
Computation of Ellipse Orientation and Semiaxes / 371
Example Problem of Standard Error Ellipse
Calculations / 376
19.3.1 Error Ellipse for Station Wisconsin / 376
19.3.2 Error Ellipse for Station Campus / 377
19.3.3 Drawing the Standard Error Ellipse / 378
19.4
Another Example Problem / 378
19.5
Error Ellipse Confidence Level / 379
19.6
Error Ellipse Advantages / 381
19.6.1 Survey Network Design / 381
19.6.2 Example Network / 383
19.7
Other Measures of Station Uncertainty / 385
Problems / 386
20
CONSTRAINT EQUATIONS
388
20.1
20.2
20.3
Introduction / 388
Adjustment of Control Station Coordinates / 388
Holding Control Station Coordinates and Directions of
Lines Fixed in a Trilateration Adjustment / 394
20.3.1 Holding the Direction of a Line Fixed by
Elimination of Constraints / 395
20.4
Helmert’s Method / 398
20.5
Redundancies in a Constrained Adjustment / 403
20.6
Enforcing Constraints through Weighting / 403
Problems / 406
21
BLUNDER DETECTION IN HORIZONTAL NETWORKS
21.1
21.2
Introduction / 409
A Priori Methods for Detecting Blunders in
Observations / 410
21.2.1 Use of the K Matrix / 410
21.2.2 Traverse Closure Checks / 411
409
CONTENTS
xiii
21.3
21.4
A Posteriori Blunder Detection / 412
Development of the Covariance Matrix for the
Residuals / 414
21.5
Detection of Outliers in Observations / 416
21.6
Techniques Used in Adjusting Control / 418
21.7
Data Set with Blunders / 420
21.8
Some Further Considerations / 428
21.8.1 Internal Reliability / 429
21.8.2 External Reliability / 429
21.9
Survey Design / 430
Problems / 432
22
GENERAL LEAST SQUARES METHOD AND ITS APPLICATION
TO CURVE FITTING AND COORDINATE
TRANSFORMATIONS
437
22.1
22.2
Introduction to General Least Squares / 437
General Least Squares Equations for Fitting a Straight
Line / 437
22.3
General Least Squares Solution / 439
22.4
Two-Dimensional Coordinate Transformation by General
Least Squares / 443
22.4.1 Two-Dimensional Conformal Coordinate
Transformation / 444
22.4.2 Two-Dimensional Affine Coordinate
Transformation / 447
22.4.3 Two-Dimensional Projective Transformation / 448
22.5
Three-Dimensional Conformal Coordinate Transformation
by General Least Squares / 449
Problems / 451
23
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT 454
23.1
23.2
Introduction / 454
Linearization of Equations / 456
23.2.1 Slant Distance Observations / 457
23.2.2 Azimuth Observations / 457
23.2.3 Vertical Angle Observations / 459
23.2.4 Horizontal Angle Observations / 459
23.2.5 Differential Leveling Observations / 460
23.2.6 Horizontal Distance Observations / 460
xiv
CONTENTS
23.3
23.4
Minimum Number of Constraints / 462
Example Adjustment / 462
23.4.1 Addition of Slant Distances / 464
23.4.2 Addition of Horizontal Angles / 465
23.4.3 Addition of Zenith Angles / 466
23.4.4 Addition of Observed Azimuths / 467
23.4.5 Addition of Elevation Differences / 467
23.4.6 Adjustment of Control Stations / 468
23.4.7 Results of Adjustment / 469
23.4.8 Updating Geodetic Coordinates / 469
23.5
Building an Adjustment / 471
23.6
Comments on Systematic Errors / 471
Problems / 474
24
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
478
24.1
24.2
24.3
24.4
Introduction / 478
Helmert Transformation / 480
Rotations between Coordinate Systems / 484
Combining GPS Baseline Vectors with Traditional
Observations / 484
24.5
Other Considerations / 489
Problems / 489
25
ANALYSIS OF ADJUSTMENTS
492
25.1
25.2
Introduction / 492
Basic Concepts, Residuals, and the Normal
Distribution / 492
25.3
Goodness-of-Fit Test / 496
25.4
Comparison of Residual Plots / 499
25.5
Use of Statistical Blunder Detection / 501
Problems / 502
26
COMPUTER OPTIMIZATION
26.1
26.2
26.3
26.4
26.5
Introduction / 504
Storage Optimization / 504
Direct Formation of the Normal Equations / 507
Cholesky Decomposition / 508
Forward and Back Solutions / 511
504
CONTENTS
xv
26.6
Using the Cholesky Factor to Find the Inverse of the
Normal Matrix / 512
26.7
Spareness and Optimization of the Normal Matrix / 513
Problems / 518
APPENDIX A
INTRODUCTION TO MATRICES
520
A.1
A.2
A.3
A.4
A.5
A.6
A.7
A.8
A.9
Introduction / 520
Definition of a Matrix / 520
Size or Dimensions of a Matrix / 521
Types of Matrices / 522
Matrix Equality / 523
Addition or Subtraction of Matrices / 524
Scalar Multiplication of a Matrix / 524
Matrix Multiplication / 525
Computer Algorithms for Matrix Operations / 528
A.9.1 Addition or Subtraction of Two
Matrices / 528
A.9.2 Matrix Multiplication / 529
A.10 Use of the MATRIX Software / 531
Problems / 531
APPENDIX B
SOLUTION OF EQUATIONS BY MATRIX METHODS 534
B.1 Introduction / 534
B.2 Inverse Matrix / 534
B.3 Inverse of a 2 ⫻ 2 Matrix / 535
B.4 Inverses by Adjoints / 537
B.5 Inverses by Row Transformations / 538
B.6 Example Problem / 542
Problems / 543
APPENDIX C
NONLINEAR EQUATIONS AND TAYLOR’S
THEOREM
C.1
C.2
C.3
C.4
Introduction / 546
Taylor Series Linearization of Nonlinear
Equations / 546
Numerical Example / 547
Using Matrices to Solve Nonlinear
Equations / 549
546
xvi
CONTENTS
C.5 Simple Matrix Example / 550
C.6 Practical Example / 551
Problems / 554
APPENDIX D
NORMAL ERROR DISTRIBUTION CURVE AND
OTHER STATISTICAL TABLES
D.1
D.2
556
Development of the Normal Distribution Curve
Equation / 556
Other Statistical Tables / 564
D.2.1 2 Distribution / 564
D.2.2 t Distribution / 566
D.2.3 F Distribution / 568
APPENDIX E
CONFIDENCE INTERVALS FOR THE MEAN
576
APPENDIX F
MAP PROJECTION COORDINATE SYSTEMS
582
F.1
F.2
F.3
F.4
APPENDIX G
Introduction / 582
Mathematics of the Lambert Conformal Conic
Map Projection / 583
F.2.1 Zone Constants / 584
F.2.2 Direct Problem / 585
F.2.3 Inverse Problem / 585
Mathematics of the Transverse Mercator / 586
F.3.1 Zone Constants / 587
F.3.2 Direct Problem / 588
F.3.3 Inverse Problem / 588
Reduction of Observations / 590
F.4.1 Reduction of Distances / 590
F.4.2 Reduction of Geodetic Azimuths / 593
COMPANION CD
G.1
G.2
G.3
Introduction / 595
File Formats and Memory Matters / 596
Software / 596
G.3.1 ADJUST / 596
G.3.2 STATS / 597
G.3.3 MATRIX / 598
G.3.4 Mathcad Worksheets / 598
595
CONTENTS
G.4
xvii
Using the Software as an Instructional
Aid / 599
BIBLIOGRAPHY
600
INDEX
603
PREFACE
No measurement is ever exact. As a corollary, every measurement contains
error. These statements are fundamental and universally accepted. It follows
logically, therefore, that surveyors, who are measurement specialists, should
have a thorough understanding of errors. They must be familiar with the
different types of errors, their sources, and their expected magnitudes. Armed
with this knowledge they will be able to (1) adopt procedures for reducing
error sizes when making their measurements and (2) account rigorously for
the presence of errors as they analyze and adjust their measured data. This
book is devoted to creating a better understanding of these topics.
In recent years, the least squares method of adjusting spatial data has been
rapidly gaining popularity as the method used for analyzing and adjusting
surveying data. This should not be surprising, because the method is the most
rigorous adjustment procedure available. It is soundly based on the mathematical theory of probability; it allows for appropriate weighting of all observations in accordance with their expected precisions; and it enables
complete statistical analyses to be made following adjustments so that the
expected precisions of adjusted quantities can be determined. Procedures for
employing the method of least squares and then statistically analyzing the
results are major topics covered in this book.
In years past, least squares was seldom used for adjusting surveying data
because the time required to set up and solve the necessary equations was
too great for hand methods. Now computers have eliminated that disadvantage. Besides advances in computer technology, some other recent developments have also led to increased use of least squares. Prominent among these
are the global positioning system (GPS) and geographic information systems
and land information systems (GISs and LISs). These systems rely heavily
xix
xx
PREFACE
on rigorous adjustment of data and statistical analysis of the results. But
perhaps the most compelling of all reasons for the recent increased interest
in least squares adjustment is that new accuracy standards for surveys are
being developed that are based on quantities obtained from least squares adjustments. Thus, surveyors of the future will not be able to test their measurements for compliance with these standards unless they adjust their data
using least squares. Clearly, modern surveyors must be able to apply the
method of least squares to adjust their measured data, and they must also be
able to perform a statistical evaluation of the results after making the
adjustments.
This book originated in 1968 as a set of lecture notes for a course taught
to a group of practicing surveyors in the San Francisco Bay area by Professor
Paul R. Wolf. The notes were subsequently bound and used as the text for
formal courses in adjustment computations taught at both the University of
California–Berkeley and the University of Wisconsin–Madison. In 1980, a
second edition was produced that incorporated many changes and suggestions
from students and others who had used the notes. The second edition, published by Landmark Enterprises, has been distributed widely to practicing
surveyors and has also been used as a textbook for adjustment computations
courses in several colleges and universities.
For the fourth edition, new chapters on the three-dimensional geodetic
network adjustments, combining GPS baseline vectors and terrestrial observations in an adjustment, the Helmert transformation, analysis of adjustments,
and state plane coordinate computations are added. These are in keeping with
the modern survey firm that collects data in three dimensions and needs to
analyze large data sets. Additionally, Chapter 4 of the third edition has been
divided into two new chapters on confidence intervals and statistical testing.
This edition has greatly expanded and modified the number of problems for
each chapter to provide readers with ample practice problems. For instructors
who adopt this book in their classes, a Solutions Manual to Accompany Adjustment Computations is also available from the publisher.
Two new appendixes have been added, including one on map projection
coordinate systems and another on the companion CD. The software included
on the CD for this book has also been greatly expanded and updated. A
Mathcad electronic book added to the companion CD demonstrates the computations for many of the example problems in the text. To obtain a greater
understanding of the material contained in this text, these electronic worksheets allow the reader to explore the intermediate computations in more
detail. For readers not having the Mathcad software package, hypertext
markup language (html) files are included on the CD for browsing.
The software STATS, ADJUST, and MATRIX are now Windows-based and
will run on a PC-compatible computer. The first package, called STATS, performs basic statistical analyses. For any given set of measured data, it will
compute the mean, median, mode, and standard deviation, and develop and
plot the histogram and normal distribution curve. The second package, called
PREFACE
xxi
ADJUST, contains programs for performing specific least-squares adjustments. Level nets, horizontal surveys (trilateration, triangulation, traverses,
and horizontal network surveys), GPS networks, and traditional threedimensional surveys can be adjusted using software in this package. It also
contains programs to compute the least-squares solution for a variety of coordinate transformations, and to determine the least squares fit of a line, parabola, or circle to a set of data points. Each of these programs computes
residuals and standard deviations following the adjustment. The third program
package, called MATRIX, performs a collection of basic matrix operations,
such as addition, subtraction, transpose, multiplication, inverse, and more.
Using this program, systems of simultaneous linear equations can be solved
quickly and conveniently, and the basic algorithm for doing least squares
adjustments can be solved in a stepwise fashion. For those who wish to develop their own software, the book provides several helpful computer algorithms in the languages of BASIC, C, FORTRAN, and PASCAL. Additionally,
the Mathcad worksheets demonstrate the use of functions in developing modular programs.
This current edition now consists of 26 chapters and several appendixes.
The chapters are arranged in the order found most convenient in teaching
college courses on adjustment computations. It is believed that this order also
best facilitates practicing surveyors who use the book for self-study. In earlier
chapters we define terms and introduce students to the fundamentals of errors
and methods for analyzing them. The next several chapters are devoted to the
subject of error propagation in the various types of traditional surveying measurements. Then chapters follow that describe observation weighting and introduce the least-squares method for adjusting observations. Application of
least squares in adjusting basic types of surveys are then presented in separate
chapters. Adjustment of level nets, trilateration, triangulation, traverses, and
horizontal networks, GPS networks, and traditional three-dimensional surveys
are included. The subject of error ellipses is covered in a separate chapter.
Procedures for applying least squares in curve fitting and in computing coordinate transformations are also presented. The more advanced topics of
blunder detection, the method of general least squares, and computer optimization are covered in the last chapters.
As with previous editions, matrix methods, which are so well adapted to
adjustment computations, continue to be used in this edition. For those students who have never studied matrices, or those who wish to review this
topic, an introduction to matrix methods is given in Appendixes A and B.
Those students who have already studied matrices can conveniently skip this
subject.
Least-squares adjustments often require the formation and solution of nonlinear equations. Procedures for linearizing nonlinear equations by Taylor’s
theorem are therefore important in adjustment computations, and this topic is
presented in Appendix C. Appendix D contains several statistical tables including the standard normal error distribution, the 2 distribution, Student’s t
xxii
PREFACE
distribution, and a set of F-distribution tables. These tables are described at
appropriate locations in the text, and their use is demonstrated with example
problems.
Basic courses in statistics and calculus are necessary prerequisites to understanding some of the theoretical coverage and equation derivations given
herein. Nevertheless, those who do not have these courses as background but
who wish to learn how to apply least squares in adjusting surveying observations can study Chapters 1 through 3, skip Chapters 4 through 8, and then
proceed with the remaining chapters.
Besides being appropriate for use as a textbook in college classes, this
book will be of value to practicing surveyors and geospatial information managers. The authors hope that through the publication of this book, least
squares adjustment and rigorous statistical analyses of surveying data will
become more commonplace, as they should.
ACKNOWLEDGMENTS
Through the years many people have contributed to the development of this
book. As noted in the preface, the book has been used in continuing education
classes taught to practicing surveyors as well as in classes taken by students
at the University of California–Berkeley, the University of Wisconsin–
Madison, and the Pennsylvania State University–Wilkes-Barre. The students
in these classes have provided data for many of the example problems and
have supplied numerous helpful suggestions for improvements throughout the
book. The authors gratefully acknowledge their contributions.
Earlier editions of the book benefited specifically from the contributions
of Mr. Joseph Dracup of the National Geodetic Survey, Professor Harold
Welch of the University of Michigan, Professor Sandor Veress of the University of Washington, Mr. Charles Schwarz of the National Geodetic Survey,
Mr. Earl Burkholder of the New Mexico State University, Dr. Herbert Stoughton of Metropolitan State College, Dr. Joshua Greenfeld of New Jersey Institute of Technology, Dr. Steve Johnson of Purdue University, Mr. Brian
Naberezny of Pennsylvania State University, and Professor David Mezera of
the University of Wisconsin–Madison. The suggestions and contributions of
these people were extremely valuable and are very much appreciated.
To improve future editions, the author will gratefully accept any constructive criticisms of this edition and suggestions for its improvement.
xxiii
CHAPTER 1
INTRODUCTION
1.1
INTRODUCTION
We currently live in what is often termed the information age. Aided by new
and emerging technologies, data are being collected at unprecedented rates in
all walks of life. For example, in the field of surveying, total station instruments, global positioning system (GPS) equipment, digital metric cameras,
and satellite imaging systems are only some of the new instruments that are
now available for rapid generation of vast quantities of measured data.
Geographic Information Systems (GISs) have evolved concurrently with
the development of these new data acquisition instruments. GISs are now
used extensively for management, planning, and design. They are being applied worldwide at all levels of government, in business and industry, by
public utilities, and in private engineering and surveying offices. Implementation of a GIS depends upon large quantities of data from a variety of
sources, many of them consisting of observations made with the new instruments, such as those noted above.
Before data can be utilized, however, whether for surveying and mapping
projects, for engineering design, or for use in a geographic information system, they must be processed. One of the most important aspects of this is to
account for the fact that no measurements are exact. That is, they always
contain errors.
The steps involved in accounting for the existence of errors in measurements consist of (1) performing statistical analyses of the observations to
assess the magnitudes of their errors and to study their distributions to determine whether or not they are within acceptable tolerances; and if the observations are acceptable, (2) adjusting them so that they conform to exact
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
1
2
INTRODUCTION
geometric conditions or other required constraints. Procedures for performing
these two steps in processing measured data are principal subjects of this
book.
1.2
DIRECT AND INDIRECT MEASUREMENTS
Measurements are defined as observations made to determine unknown quantities. They may be classified as either direct or indirect. Direct measurements
are made by applying an instrument directly to the unknown quantity and
observing its value, usually by reading it directly from graduated scales on
the device. Determining the distance between two points by making a direct
measurement using a graduated tape, or measuring an angle by making a
direct observation from the graduated circle of a theodolite or total station
instrument, are examples of direct measurements.
Indirect measurements are obtained when it is not possible or practical to
make direct measurements. In such cases the quantity desired is determined
from its mathematical relationship to direct measurements. Surveyors may,
for example, measure angles and lengths of lines between points directly and
use these measurements to compute station coordinates. From these coordinate values, other distances and angles that were not measured directly may
be derived indirectly by computation. During this procedure, the errors that
were present in the original direct observations are propagated (distributed)
by the computational process into the indirect values. Thus, the indirect measurements (computed station coordinates, distances, and angles) contain errors
that are functions of the original errors. This distribution of errors is known
as error propagation. The analysis of how errors propagate is also a principal
topic of this book.
1.3
MEASUREMENT ERROR SOURCES
It can be stated unconditionally that (1) no measurement is exact, (2) every
measurement contains errors, (3) the true value of a measurement is never
known, and thus (4) the exact sizes of the errors present are always unknown.
These facts can be illustrated by the following. If an angle is measured with
a scale divided into degrees, its value can be read only to perhaps the nearest
tenth of a degree. If a better scale graduated in minutes were available and
read under magnification, however, the same angle might be estimated to
tenths of a minute. With a scale graduated in seconds, a reading to the nearest
tenth of a second might be possible. From the foregoing it should be clear
that no matter how well the observation is taken, a better one may be possible.
Obviously, in this example, observational accuracy depends on the division
size of the scale. But accuracy depends on many other factors, including the
overall reliability and refinement of the equipment used, environmental con-
1.4
DEFINITIONS
3
ditions that exist when the observations are taken, and human limitations (e.g.,
the ability to estimate fractions of a scale division). As better equipment is
developed, environmental conditions improve, and observer ability increases,
observations will approach their true values more closely, but they can never
be exact.
By definition, an error is the difference between a measured value for any
quantity and its true value, or
ε⫽y⫺
(1.1)
where ε is the error in an observation, y the measured value, and its true
value.
As discussed above, errors stem from three sources, which are classified
as instrumental, natural, and personal:
1. Instrumental errors. These errors are caused by imperfections in instrument construction or adjustment. For example, the divisions on a
theodolite or total station instrument may not be spaced uniformly.
These error sources are present whether the equipment is read manually
or digitally.
2. Natural errors. These errors are caused by changing conditions in the
surrounding environment, including variations in atmospheric pressure,
temperature, wind, gravitational fields, and magnetic fields.
3. Personal errors. These errors arise due to limitations in human senses,
such as the ability to read a micrometer or to center a level bubble. The
sizes of these errors are affected by the personal ability to see and by
manual dexterity. These factors may be influenced further by temperature, insects, and other physical conditions that cause humans to behave
in a less precise manner than they would under ideal conditions.
1.4
DEFINITIONS
From the discussion thus far, it can be stated with absolute certainty that all
measured values contain errors, whether due to lack of refinement in readings,
instabilities in environmental conditions, instrumental imperfections, or human limitations. Some of these errors result from physical conditions that
cause them to occur in a systematic way, whereas others occur with apparent
randomness. Accordingly, errors are classified as either systematic or random.
But before defining systematic and random errors, it is helpful to define
mistakes.
1. Mistakes. These are caused by confusion or by an observer’s carelessness. They are not classified as errors and must be removed from any
4
INTRODUCTION
set of observations. Examples of mistakes include (a) forgetting to set
the proper parts per million (ppm) correction on an EDM instrument,
or failure to read the correct air temperature, (b) mistakes in reading
graduated scales, and (c) blunders in recording (i.e., writing down 27.55
for 25.75). Mistakes are also known as blunders or gross errors.
2. Systematic errors. These errors follow some physical law, and thus these
errors can be predicted. Some systematic errors are removed by following correct measurement procedures (e.g., balancing backsight and foresight distances in differential leveling to compensate for Earth curvature
and refraction). Others are removed by deriving corrections based on
the physical conditions that were responsible for their creation (e.g.,
applying a computed correction for Earth curvature and refraction on a
trigonometric leveling observation). Additional examples of systematic
errors are (a) temperature not being standard while taping, (b) an index
error of the vertical circle of a theodolite or total station instrument,
and (c) use of a level rod that is not of standard length. Corrections for
systematic errors can be computed and applied to observations to eliminate their effects. Systematic errors are also known as biases.
3. Random errors. These are the errors that remain after all mistakes and
systematic errors have been removed from the measured values. In general, they are the result of human and instrument imperfections. They
are generally small and are as likely to be negative as positive. They
usually do not follow any physical law and therefore must be dealt with
according to the mathematical laws of probability. Examples of random
errors are (a) imperfect centering over a point during distance measurement with an EDM instrument, (b) bubble not centered at the instant a
level rod is read, and (c) small errors in reading graduated scales. It is
impossible to avoid random errors in measurements entirely. Although
they are often called accidental errors, their occurrence should not be
considered an accident.
1.5
PRECISION VERSUS ACCURACY
Due to errors, repeated observation of the same quantity will often yield
different values. A discrepancy is defined as the algebraic difference between
two observations of the same quantity. When small discrepancies exist between repeated observations, it is generally believed that only small errors
exist. Thus, the tendency is to give higher credibility to such data and to call
the observations precise. However, precise values are not necessarily accurate
values. To help understand the difference between precision and accuracy, the
following definitions are given:
1. Precision is the degree of consistency between observations based on
the sizes of the discrepancies in a data set. The degree of precision
1.5
PRECISION VERSUS ACCURACY
5
attainable is dependent on the stability of the environment during the
time of measurement, the quality of the equipment used to make the
observations, and the observer’s skill with the equipment and observational procedures.
2. Accuracy is the measure of the absolute nearness of a measured quantity
to its true value. Since the true value of a quantity can never be determined, accuracy is always an unknown.
The difference between precision and accuracy can be demonstrated using
distance observations. Assume that the distance between two points is paced,
taped, and measured electronically and that each procedure is repeated five
times. The resulting observations are:
Observation
Pacing,
p
Taping,
t
EDM,
e
1
2
3
4
5
571
563
566
588
557
567.17
567.08
567.12
567.38
567.01
567.133
567.124
567.129
567.165
567.114
The arithmetic means for these data sets are 569, 567.15, and 567.133,
respectively. A line plot illustrating relative values of the electronically measured distances denoted by e, and the taped distances, denoted by t, is shown
in Figure 1.1. Notice that although the means of the EDM data set and of the
taped observations are relatively close, the EDM set has smaller discrepancies.
This indicates that the EDM instrument produced a higher precision. However, this higher precision does not necessarily prove that the mean of the
electronically measured data set is implicitly more accurate than the mean of
the taped values. In fact, the opposite may be true if the reflector constant
was entered incorrectly causing a large systematic error to be present in all
the electronically measured distances. Because of the larger discrepancies, it
is unlikely that the mean of the paced distances is as accurate as either of the
other two values. But its mean could be more accurate if large systematic
errors were present in both the taped and electronically measured distances.
Figure 1.1 Line plot of distance quantities.
6
INTRODUCTION
Another illustration explaining differences between precision and accuracy
involves target shooting, depicted in Figure 1.2. As shown, four situations
can occur. If accuracy is considered as closeness of shots to the center of a
target at which a marksman shoots, and precision as the closeness of the shots
to each other, (1) the data may be both precise and accurate, as shown in
Figure 1.2(a); (2) the data may produce an accurate mean but not be precise,
as shown in Figure 1.2(b); (3) the data may be precise but not accurate, as
shown in Figure 1.2(c); or (4) the data may be neither precise nor accurate,
as shown in Figure 1.2(d).
Figure 1.2(a) is the desired result when observing quantities. The other
cases can be attributed to the following situations. The results shown in Figure
1.2(b) occur when there is little refinement in the observational process.
Someone skilled at pacing may achieve these results. Figure 1.2(c) generally
occurs when systematic errors are present in the observational process. For
example, this can occur in taping if corrections are not made for tape length
and temperature, or with electronic distance measurements when using the
wrong combined instrument–reflector constant. Figure 1.2(d) shows results
obtained when the observations are not corrected for systematic errors and
are taken carelessly by the observer (or the observer is unskilled at the particular measurement procedure).
In general, when making measurements, data such as those shown in Figure
1.2(b) and (d) are undesirable. Rather, results similar to those shown in Figure
1.2(a) are preferred. However, in making measurements the results of Figure
1.2(c) can be just as acceptable if proper steps are taken to correct for the
presence of systematic errors. (This correction would be equivalent to a
Figure 1.2 Examples of precision versus accuracy.
1.6
REDUNDANT MEASUREMENTS IN SURVEYING AND THEIR ADJUSTMENT
7
marksman realigning the sights after taking shots.) To make these corrections,
(1) the specific types of systematic errors that have occurred in the observations must be known, and (2) the procedures used in correcting them must
be understood.
1.6 REDUNDANT MEASUREMENTS IN SURVEYING AND
THEIR ADJUSTMENT
As noted earlier, errors exist in all observations. In surveying, the presence
of errors is obvious in many situations where the observations must meet
certain conditions. For example, in level loops that begin and close on the
same benchmark, the elevation difference for the loop must equal zero. However, in practice, this is hardly ever the case, due to the presence of random
errors. (For this discussion it is assumed that all mistakes have been eliminated from the observations and appropriate corrections have been applied to
remove all systematic errors.) Other conditions that disclose errors in surveying observations are that (1) the three measured angles in a plane triangle
must total 180⬚, (2) the sum of the angles measured around the horizon at
any point must equal 360⬚, and (3) the algebraic sum of the latitudes (and
departures) must equal zero for closed polygon traverses that begin and end
on the same station. Many other conditions could be cited; however, in any
of them, the observations rarely, if ever, meet the required conditions, due to
the presence of random errors.
The examples above not only demonstrate that errors are present in surveying observations but also the importance of redundant observations; those
measurements made that are in excess of the minimum number that are
needed to determine the unknowns. For example, two measurements of the
length of a line yield one redundant observation. The first observation would
be sufficient to determine the unknown length, and the second is redundant.
However, this second observation is very valuable. First, by examining the
discrepancy between the two values, an assessment of the size of the error in
the observations can be made. If a large discrepancy exists, a blunder or large
error is likely to have occurred. In that case, measurements of the line would
be repeated until two values having an acceptably small discrepancy were
obtained. Second, the redundant observation permits an adjustment to be made
to obtain a final value for the unknown line length, and that final adjusted
value will be more precise statistically than either of the individual observations. In this case, if the two observations were of equal precision, the adjusted
value would be the simple mean.
Each of the specific conditions cited in the first paragraph of this section
involves one redundant observation. For example, there is one redundant observation when the three angles of a plane triangle are observed. This is true
because with two observed angles, say A and B, the third could be computed
as C ⫽ 180⬚ ⫺ A ⫺ B, and thus observation of C is unnecessary. However,
8
INTRODUCTION
measuring angle C enables an assessment of the errors in the angles and also
makes an adjustment possible to obtain final angles with statistically improved
precisions. Assuming that the angles were of equal precision, the adjustment
would enforce a 180⬚ sum for the three angles by distributing the total discrepancy in equal parts to each angle.
Although the examples cited here are indeed simple, they help to define
redundant measurements and to illustrate their importance. In large surveying
networks, the number of redundant observations can become extremely large,
and the adjustment process is somewhat more involved than it is for the simple
examples given here.
Prudent surveyors always make redundant observations in their work, for
the two important reasons indicated above: (1) to make it possible to assess
errors and make decisions regarding acceptance or rejection of observations,
and (2) to make possible an adjustment whereby final values with higher
precisions are determined for the unknowns.
1.7
ADVANTAGES OF LEAST SQUARES ADJUSTMENT
As indicated in Section 1.6, in surveying it is recommended that redundant
observations always be made and that adjustments of the observations always
be performed. These adjustments account for the presence of errors in the
observations and increase the precision of the final values computed for the
unknowns. When an adjustment is completed, all observations are corrected
so that they are consistent throughout the survey network [i.e., the same values
for the unknowns are determined no matter which corrected observation(s)
are used to compute them].
Many different methods have been derived for making adjustments in surveying; however, the method of least squares should be used because it has
significant advantages over all other rule-of-thumb procedures. The advantages of least squares over other methods can be summarized with the following four general statements: (1) it is the most rigorous of adjustments; (2)
it can be applied with greater ease than other adjustments; (3) it enables
rigorous postadjustment analyses to be made; and (4) it can be used to perform presurvey planning. These advantages are discussed further below.
Least squares adjustment is rigorously based on the theory of mathematical
probability, whereas in general, the other methods do not have this rigorous
base. As described later in the book, in a least squares adjustment, the following condition of mathematical probability is enforced: The sum of the
squares of the errors times their respective weights is minimized. By enforcing
this condition in any adjustment, the set of errors that is computed has the
highest probability of occurrence. Another aspect of least squares adjustment
that adds to its rigor is that it permits all observations, regardless of their
number or type, to be entered into the adjustment and used simultaneously
in the computations. Thus, an adjustment can combine distances, horizontal
1.7
ADVANTAGES OF LEAST SQUARES ADJUSTMENT
9
angles, azimuths, zenith or vertical angles, height differences, coordinates,
and even GPS observations. One important additional asset of least squares
adjustment is that it enables ‘‘relative weights’’ to be applied to the observations in accordance with their estimated relative reliabilities. These reliabilities are based on estimated precisions. Thus, if distances were observed
in the same survey by pacing, taping, and using an EDM instrument, they
could all be combined in an adjustment by assigning appropriate relative
weights.
Years ago, because of the comparatively heavy computational effort involved in least squares, nonrigorous or ‘‘rule-of-thumb’’ adjustments were
most often used. However, now because computers have eliminated the computing problem, the reverse is true and least squares adjustments are performed more easily than these rule-of-thumb techniques. Least squares
adjustments are less complicated because the same fundamental principles are
followed regardless of the type of survey or the type of observations. Also,
the same basic procedures are used regardless of the geometric figures involved (e.g., triangles, closed polygons, quadrilaterals, or more complicated
networks). On the other hand, rules of thumb are not the same for all types
of surveys (e.g., level nets use one rule and traverses use another), and they
vary for different geometric shapes. Furthermore, the rule of thumb applied
for a particular survey by one surveyor may be different from that applied by
another surveyor. A favorable characteristic of least squares adjustments is
that there is only one rigorous approach to the procedure, and thus no matter
who performs the adjustment for any particular survey, the same results will
be obtained.
Least squares has the advantage that after an adjustment has been finished,
a complete statistical analysis can be made of the results. Based on the sizes
and distribution of the errors, various tests can be conducted to determine if
a survey meets acceptable tolerances or whether the observations must be
repeated. If blunders exist in the data, these can be detected and eliminated.
Least squares enables precisions for the adjusted quantities to be determined
easily and these precisions can be expressed in terms of error ellipses for
clear and lucid depiction. Procedures for accomplishing these tasks are described in subsequent chapters.
Besides its advantages in adjusting survey data, least squares can be used
to plan surveys. In this application, prior to conducting a needed survey,
simulated surveys can be run in a trial-and-error procedure. For any project,
an initial trial geometric figure for the survey is selected. Based on the figure,
trial observations are either computed or scaled. Relative weights are assigned
to the observations in accordance with the precision that can be estimated
using different combinations of equipment and field procedures. A least
squares adjustment of this initial network is then performed and the results
analyzed. If goals have not been met, the geometry of the figure and the
observation precisions are varied and the adjustment performed again. In this
process different types of observations can be used, and observations can be
10
INTRODUCTION
added or deleted. These different combinations of geometric figures and observations are varied until one is achieved that produces either optimum or
satisfactory results. The survey crew can then proceed to the field, confident
that if the project is conducted according to the design, satisfactory results
will be obtained. This technique of applying least squares in survey planning
is discussed in later chapters.
1.8 OVERVIEW OF THE BOOK
In the remainder of the book the interrelationship between observational errors
and their adjustment is explored. In Chapters 2 through 5, methods used to
determine the reliability of observations are described. In these chapters, the
ways that errors of multiple observations tend to be distributed are illustrated,
and techniques used to compare the quality of different sets of measured
values are examined. In Chapters 6 through 9 and in Chapter 13, methods
used to model error propagation in observed and computed quantities are
discussed. In particular, error sources present in traditional surveying techniques are examined, and the ways in which these errors propagate throughout
the observational and computational processes are explained. In the remainder
of the book, the principles of least squares are applied to adjust observations
in accordance with random error theory and techniques used to locate mistakes in observations are examined.
PROBLEMS
1.1
Describe an example in which directly measured quantities are used to
obtain an indirect measurement.
1.2
Identify the direct and indirect measurements used in computing traverse station coordinates.
1.3
Explain the difference between systematic and random errors.
1.4
List possible systematic and random errors when measuring:
(a) a distance with a tape.
(b) a distance with an EDM.
(c) an angle with a total station.
(d) the difference in elevation using an automatic level.
1.5
List three examples of mistakes that can be made when measuring an
angle with total station instruments.
PROBLEMS
11
1.6
Identify each of the following errors as either systematic or random.
(a) Reading a level rod.
(b) Not holding a level rod plumb.
(c) Leveling an automatic leveling instrument.
(d) Using a level rod that has one foot removed from the bottom of
the rod.
1.7
In your own words, define the difference between precision and
accuracy.
1.8
Identify each of the following errors according to its source (natural,
instrumental, personal):
(a) Level rod length.
(b) EDM–reflector constant.
(c) Air temperature in an EDM observation.
(d) Reading a graduation on a level rod.
(e) Earth curvature in leveling observations.
(f) Horizontal collimation error of an automatic level.
1.9
The calibrated length of a particular line is 400.012 m. A length of
400.015 m is obtained using an EDM. What is the error in the
observation?
1.10 In Problem 1.9, if the length observed is 400.007 m, what is the error
in the observation?
1.11 Why do surveyors measure angles using both faces of a total station
(i.e., direct and reversed)?
1.12 Give an example of compensating systematic errors in a vertical angle
observation when the angle is measured using both faces of the
instrument.
1.13 What systematic errors exist in taping the length of a line?
1.14 Discuss the importance of making redundant observations in surveying.
1.15 List the advantages of making adjustments by the method of least
squares.
CHAPTER 2
OBSERVATIONS AND THEIR ANALYSIS
2.1
INTRODUCTION
Sets of data can be represented and analyzed using either graphical or numerical methods. Simple graphical analyses to depict trends commonly appear
in newspapers and on television. A plot of the daily variation of the closing
Dow Jones industrial average over the past year is an example. A bar chart
showing daily high temperatures over the past month is another. Data can
also be presented in numerical form and be subjected to numerical analysis.
As a simple example, instead of using a bar chart, the daily high temperatures
could be tabulated and the mean computed. In surveying, observational data
can also be represented and analyzed either graphically or numerically. In
this chapter some rudimentary methods for doing so are explored.
2.2
SAMPLE VERSUS POPULATION
Due to time and financial constraints, generally only a small data sample is
collected from a much larger, possibly infinite population. For example, political parties may wish to know the percentage of voters who support their
candidate. It would be prohibitively expensive to query the entire voting population to obtain the desired information. Instead, polling agencies select a
sample subset of voters from the voting population. This is an example of
population sampling.
As another example, suppose that an employer wishes to determine the
relative measuring capabilities of two prospective new employees. The can12
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
2.3
RANGE AND MEDIAN
13
didates could theoretically spend days or even weeks demonstrating their
abilities. Obviously, this would not be very practical, so instead, the employer
could have each person record a sample of readings and from the readings
predict the person’s abilities. The employer could, for instance, have each
candidate read a micrometer 30 times. The 30 readings would represent a
sample of the entire population of possible readings. In fact, in surveying,
every time that distances, angles, or elevation differences are measured, samples are being collected from a population of measurements.
From the preceding discussion, the following definitions can be made:
1. Population. A population consists of all possible measurements that can
be made on a particular item or procedure. Often, a population has an
infinite number of data elements.
2. Sample. A sample is a subset of data selected from the population.
2.3
RANGE AND MEDIAN
Suppose that a 1-second (1ⴖ) micrometer theodolite is used to read a direction
50 times. The seconds portions of the readings are shown in Table 2.1. These
readings constitute what is called a data set. How can these data be organized
to make them more meaningful? How can one answer the question: Are the
data representative of readings that should reasonably be expected with this
instrument and a competent operator? What statistical tools can be used to
represent and analyze this data set?
One quick numerical method used to analyze data is to compute its range,
also called dispersion. A range is the difference between the highest and
lowest values. It provides an indication of the precision of the data. From
Table 2.1, the lowest value is 20.1 and the highest is 26.1. Thus, the range is
26.1–20.1, or 6.0. The range for this data set can be compared with ranges
of other sets, but this comparison has little value when the two sets differ in
TABLE 2.1 Fifty Readings
22.7
22.3
25.5
23.8
22.9
22.2
21.9
26.1
22.6
21.7
25.4
24.2
24.7
24.4
23.4
23.3
24.3
21.2
25.3
23.9
24.0
24.8
23.2
23.7
25.9
24.6
23.8
23.0
25.0
22.3
20.5
23.5
22.0
24.1
23.1
24.1
23.1
25.9
22.8
25.3
22.5
22.9
23.8
22.6
21.8
23.2
25.2
22.8
23.6
20.1
14
OBSERVATIONS AND THEIR ANALYSIS
size. For instance, would a set of 100 data points with a range of 8.5 be better
than the set in Table 2.1? Clearly, other methods of analyzing data sets statistically would be useful.
To assist in analyzing data, it is often helpful to list the values in order of
increasing size. This was done with the data of Table 2.1 to produce the results
shown in Table 2.2. By looking at this ordered set, it is possible to determine
quickly the data’s middle value or midpoint. In this example it lies between
the values of 23.4 and 23.5. The midpoint value is also known as the median.
Since there are an even number of values in this example, the median is given
by the average of the two values closest to (which straddle) the midpoint.
That is, the median is assigned the average of the 25th and 26th entries in
the ordered set of 50 values, and thus for the data set of Table 2.2, the median
is the average of 23.4 and 23.5, or 23.45.
2.4
GRAPHICAL REPRESENTATION OF DATA
Although an ordered numerical tabulation of data allows for some data distribution analysis, it can be improved with a frequency histogram, usually
called simply a histogram. Histograms are bar graphs that show the frequency
distributions in data. To create a histogram, the data are divided into classes.
These are subregions of data that usually have a uniform range in values, or
class width. Although there are no universally applicable rules for the selection of class width, generally 5 to 20 classes are used. As a rule of thumb, a
data set of 30 values may have only five or six classes, whereas a data set of
100 values may have 10 or more classes. In general, the smaller the data set,
the lower the number of classes used.
The histogram class width (range of data represented by each histogram
bar) is determined by dividing the total range by the selected number of
classes. Consider, for example, the data of Table 2.2. If they were divided
into seven classes, the class width would be the range divided by the number
of classes, or 6.0/7 ⫽ 0.857 ⬇ 0.86. The first class interval is found by
TABLE 2.2 Arranged Data
20.1
21.9
22.5
22.8
23.1
23.5
23.8
24.2
24.8
25.4
20.5
22.0
22.6
22.9
23.2
23.6
23.9
24.3
25.0
25.5
21.2
22.2
22.6
22.9
23.2
23.7
24.0
24.4
25.2
25.9
21.7
22.3
22.7
23.0
23.3
23.8
24.1
24.6
25.3
25.9
21.8
22.3
22.8
23.1
23.4
23.8
24.1
24.7
25.3
26.1
2.4
GRAPHICAL REPRESENTATION OF DATA
15
adding the class width to the lowest data value. For the data in Table 2.2, the
first class interval is from 20.1 to (20.1 ⫹ 0.86), or 20.96. This class interval
includes all data from 20.1 up to (but not including) 20.96. The next class
interval is from 20.96 up to (20.96 ⫹ 0.86), or 21.82. Remaining class intervals are found by adding the class width to the upper boundary value of the
preceding class. The class intervals for the data of Table 2.2 are listed in
column (1) of Table 2.3.
After creating class intervals, the number of data values in each interval,
called the class frequency, is tallied. Obviously, having data ordered consecutively as shown in Table 2.2 aids greatly in this counting process. Column
(2) of Table 2.3 shows the class frequency for each class interval of the data
in Table 2.2.
Often, it is also useful to calculate the class relative frequency for each
interval. This is found by dividing the class frequency by the total number of
observations. For the data in Table 2.2, the class relative frequency for the
first class interval is 2/50 ⫽ 0.04. Similarly, the class relative frequency of
the fourth interval (from 22.67 to 23.53) is 13/50 ⫽ 0.26. The class relative
frequencies for the data of Table 2.2 are given in column (3) of Table 2.3.
Notice that the sum of all class relative frequencies is always 1. The class
relative frequency enables easy determination of percentages. For instance,
the class interval from 21.82 to 22.67 contains 16% (0.16 ⫻ 100%) of the
sample observations.
A histogram is a bar graph plotted with either class frequencies or relative
class frequencies on the ordinate, versus values of the class interval bounds
on the abscissa. Using the data from Table 2.3, the histogram shown in Figure
2.1 was constructed. Notice that in this figure, relative frequencies have been
plotted as ordinates.
Histograms drawn with the same ordinate and abscissa scales can be used
to compare two data sets. If one data set is more precise than the other, it
will have comparatively tall bars in the center of the histogram, with relatively
TABLE 2.3 Frequency Count
(1)
Class
Interval
(2)
Class
Frequency
20.10–20.96
20.96–21.82
21.82–22.67
22.67–23.53
23.53–24.38
24.38–25.24
25.24–26.10
2
3
8
13
11
6
7
(3)
Class Relative
Frequency
2 / 50
3 / 50
8 / 50
13 / 50
11 / 50
6 / 50
7 / 50
50 / 50
0.04
0.06
0.16
0.26
0.22
0.12
0.14
⫽1
⫽
⫽
⫽
⫽
⫽
⫽
⫽
16
OBSERVATIONS AND THEIR ANALYSIS
Figure 2.1 Frequency histogram.
short bars near its edges. Conversely, the less precise data set will yield a
wider range of abscissa values, with shorter bars at the center.
A summary of items easily seen on a histogram include:
•
•
•
•
Whether the data are symmetrical about a central value
The range or dispersion in the measured values
The frequency of occurrence of the measured values
The steepness of the histogram, which is an indication of measurement
precision
Figure 2.2 shows several possible histogram shapes. Figure 2.2(a) depicts
a histogram that is symmetric about its central value with a single peak in
the middle. Figure 2.2(b) is also symmetric about the center but has a steeper
slope than Figure 2.2(a), with a higher peak for its central value. Assuming
Figure 2.2 Common histogram shapes.
2.6
MEASURES OF CENTRAL TENDENCY
17
the ordinate and abscissa scales to be equal, the data used to plot Figure
2.2(b) are more precise than those used for Figure 2.2(a). Symmetric histogram shapes are common in surveying practice as well as in many other fields.
In fact, they are so common that the shapes are said to be examples of a
normal distribution. In Chapter 3, reasons why these shapes are so common
are discussed.
Figure 2.2(c) has two peaks and is said to be a bimodal histogram. In the
histogram of Figure 2.2(d), there is a single peak with a long tail to the left.
This results from a skewed data set, and in particular, these data are said to
be skewed to the right. The data of histogram Figure 2.2(e) are skewed to the
left.
In surveying, the varying histogram shapes just described result from variations in personnel, physical conditions, and equipment: for example, repeated observations of a long distance made with an EDM instrument and by
taping. An EDM would probably produce data having a very narrow range,
and thus the resulting histogram would be narrow and steep with a tall central
bar such as that in Figure 2.2(b). The histogram of the same distance measured by tape and plotted at the same scales would probably be wider, with
the sides not as steep nor the central value as great, as shown in Figure 2.2(a).
Since observations in surveying practice tend to be normally distributed, bimodal or skewed histograms from measured data are not expected. The appearance of such a histogram should lead to an investigation for the cause of
this shape. For instance, if a data set from an EDM calibration plots as a
bimodal histogram, it could raise questions about whether the instrument or
reflector were moved during the measuring process, or if atmospheric conditions changed dramatically during the session. Similarly, a skewed histogram in EDM work may indicate the appearance of a weather front that
stabilized over time. The existence of multipath errors in GPS observations
could also produce these types of histogram plots.
2.5
NUMERICAL METHODS OF DESCRIBING DATA
Numerical descriptors are values computed from a data set that are used to
interpret its precision or quality. Numerical descriptors fall into three categories: (1) measures of central tendency, (2) measures of data variation, and
(3) measures of relative standing. These categories are all called statistics.
Simply described, a statistic is a numerical descriptor computed from sample
data.
2.6
MEASURES OF CENTRAL TENDENCY
Measures of central tendency are computed statistical quantities that give an
indication of the value within a data set that tends to exist at the center. The
18
OBSERVATIONS AND THEIR ANALYSIS
arithmetic mean, the median, and the mode are three such measures. They
are described as follows:
1. Arithmetic mean. For a set of n observations, y1, y2, . . . , yn, this is the
average of the observations. Its value, y, is computed from the equation
y⫽
兺ni⫽1 yi
n
(2.1)
Typically, the symbol y is used to represent a sample’s arithmetic mean
and the symbol is used to represent the population mean. Otherwise,
the same equation applies. Using Equation (2.1), the mean of the observations in Table 2.2 is 23.5.
2. Median. As mentioned previously, this is the midpoint of a sample set
when arranged in ascending or descending order. One-half of the data
are above the median and one-half are below it. When there are an odd
number of quantities, only one such value satisfies this condition. For
a data set with an even number of quantities, the average of the two
observations that straddle the midpoint is used to represent the median.
3. Mode. Within a sample of data, the mode is the most frequently occurring value. It is seldom used in surveying because of the relatively small
number of values observed in a typical set of observations. In small
sample sets, several different values may occur with the same frequency,
and hence the mode can be meaningless as a measure of central tendency. The mode for the data in Table 2.2 is 23.8.
2.7
ADDITIONAL DEFINITIONS
Several other terms, which are pertinent to the study of observations and their
analysis, are listed and defined below.
1. True value, : a quantity’s theoretically correct or exact value. As noted
in Section 1.3, the true value can never be determined.
2. Error, ε: the difference between a measured quantity and its true value.
The true value is simply the population’s arithmetic mean. Since the
true value of a measured quantity is indeterminate, errors are also indeterminate and are therefore only theoretical quantities. As given in
Equation (1.1), repeated for convenience here, errors are expressed as
εi ⫽ yi ⫺
(2.2)
where yi is the individual observation associated with εi and is the
true value for that quantity.
2.7
ADDITIONAL DEFINITIONS
19
3. Most probable value, y: that value for a measured quantity which, based
on the observations, has the highest probability of occurrence. It is
derived from a sample set of data rather than the population and is
simply the mean if the repeated measurements have the same precision.
4. Residual, v: The difference between any individual measured quantity
and the most probable value for that quantity. Residuals are the values
that are used in adjustment computations since most probable values
can be determined. The term error is frequently used when residual is
meant, and although they are very similar and behave in the same manner, there is this theoretical distinction. The mathematical expression for
a residual is
vi ⫽ y ⫺ yi
(2.3)
where vi is the residual in the ith observation, yi, and y is the most
probable value for the unknown.
5. Degrees of freedom: the number of observations that are in excess of
the number necessary to solve for the unknowns. In other words, the
number of degrees of freedom equals the number of redundant observations (see Section 1.6). As an example, if a distance between two
points is measured three times, one observation would determine the
unknown distance and the other two would be redundant. These redundant observations reveal the discrepancies and inconsistencies in observed values. This, in turn, makes possible the practice of adjustment
computations for obtaining the most probable values based on the measured quantities.
6. Variance, 2: a value by which the precision for a set of data is given.
Population variance applies to a data set consisting of an entire population. It is the mean of the squares of the errors and is given by
2 ⫽
兺ni⫽1 ε2i
n
(2.4)
Sample variance applies to a sample set of data. It is an unbiased
estimate for the population variance given in Equation (2.4) and is calculated as
S2 ⫽
兺ni⫽1 vi2
n⫺1
(2.5)
Note that Equations (2.4) and (2.5) are identical except that ε has been
changed to v and n has been changed to n ⫺ 1 in Equation (2.5). The validity
of these modifications is demonstrated in Section 2.10.
20
OBSERVATIONS AND THEIR ANALYSIS
It is important to note that the simple algebraic average of all errors in a
data set cannot be used as a meaningful precision indicator. This is because
random errors are as likely to be positive as negative, and thus the algebraic
average will equal zero. This fact is shown for a population of data in the
following simple proof. Summing Equation (2.2) for n samples gives
冘 ε ⫽ 冘 (y ⫺ ) ⫽ 冘 y ⫺ 冘 y ⫺ n
n
n
n
2
i
i
i⫽1
n
i
i⫽1
i
i⫽1
(a)
i⫽1
Then substituting Equation (2.1) into Equation (a) yields
冘ε ⫽冘y ⫺n兺n y ⫽冘y ⫺冘y ⫽0
n
i⫽1
n
i⫽1
n
i
n
n
i
i
i
i⫽1
i⫽1
i
(b)
i⫽1
Similarly, it can be shown that the mean of all residuals of a sample data set
equals zero.
7. Standard error, : the square root of the population variance. From
Equation (2.4) and this definition, the following equation is written for
the standard error:
⫽
兺ni⫽1 ε2i
冪
n
(2.6)
n
where n is the number of observations and 兺i⫽1
εi2 is the sum of the
squares of the errors. Note that both the population variance, 2, and
the standard error, , are indeterminate because true values, and hence
errors, are indeterminate.
As explained in Section 3.5, 68.3% of all observations in a population
data set lie within of the true value, . Thus, the larger the standard
error, the more dispersed are the values in the data set and the less
precise is the measurement.
8. Standard deviation, S: the square root of the sample variance. It is
calculated using the expression
S⫽
兺ni⫽1 v2i
冪n ⫺ 1
(2.7)
where S is the standard deviation, n ⫺ 1 the degrees of freedom or
n
number of redundancies, and 兺i⫽1
vi2 the sum of the squares of the
residuals. Standard deviation is an estimate for the standard error of the
population. Since the standard error cannot be determined, the standard
deviation is a practical expression for the precision of a sample set of
2.8
ALTERNATIVE FORMULA FOR DETERMINING VARIANCE
21
data. Residuals are used rather than errors because they can be calculated from most probable values, whereas errors cannot be determined.
Again, as discussed in Section 3.5, for a sample data set, 68.3% of the
observations will theoretically lie between the most probable value plus
and minus the standard deviation, S. The meaning of this statement will
be clarified in an example that follows.
9. Standard deviation of the mean: the error in the mean computed from
a sample set of measured values that results because all measured values
contain errors. The standard deviation of the mean is computed from
the sample standard deviation according to the equation
Sy ⫽
S
兹n
(2.8)
Notice that as n → ⬁, then Sy → 0. This illustrates that as the size of
the sample set approaches the total population, the computed mean y
will approach the true mean . This equation is derived in Chapter 4.
2.8
ALTERNATIVE FORMULA FOR DETERMINING VARIANCE
From the definition of residuals, Equation (2.5) is rewritten as
S2 ⫽
兺ni⫽1 (y ⫺ yi)2
(2.9)
n⫺1
Expanding Equation (2.9) yields
S2 ⫽
1
n⫺1
[(y ⫺ y1)2 ⫹ (y ⫺ y2)2 ⫹ ⫹ (y ⫺ yn)2]
(c)
Substituting Equation (2.1) for y into Equation (c) and dropping the bounds
for the summation yields
S2 ⫽
1
n⫺1
冋冉
兺 yi
n
冊 冉
2
⫺ y1
⫹
兺 yi
Expanding Equation (d) gives us
n
冊
2
⫺ y2
⫹⫹
冉
兺 yi
n
冊冥
2
⫺ yn
(d)
22
OBSERVATIONS AND THEIR ANALYSIS
S ⫽
冋冉 冊
冉 冊
兺 yi
1
2
n⫺1
2
⫺ 2y1
n
兺 yi
⫹ y22 ⫹ ⫹
兺 yi
2
1
⫹y ⫹
n
2
n
⫺ 2yn
兺 yi
n
冉 冊
册
兺 yi
n
2
⫺ 2y2
兺 yi
n
⫹ y2n
(e)
Rearranging Equation (e) and recognizing that (兺 yi /n)2 occurs n times in
Equation (e) yields
S2 ⫽
1
n⫺1
冋冉 冊
n
兺 yi
n
2
⫺2
兺 yi
n
(y1 ⫹ y2 ⫹ ⫹ yn) ⫹ y12 ⫹ y22 ⫹ ⫹ yn2
册
(ƒ)
Adding the summation symbol to Equation (ƒ) yields
S2 ⫽
1
n⫺1
冋 冉 冊 冉冘 冊
兺 yi
n
n
2
2
n
⫺
2
yi
⫹
冘y册
2
i
(g)
Factoring and regrouping similar summations in Equation (g) produces
S2 ⫽
2 1
1
1
y ⫺ 冉 ⫺ 冊冉冘 y 冊 册 ⫽
y ⫺ 冉冘 y 冊 册
冘
冘
冋
冋
n⫺1
n n
n⫺1
n
1
2
2
i
2
2
i
i
i
(h)
Multiplying the last term in Equation (h) by n/n yields
S2 ⫽
兺y
y ⫺ n冉
冘
冋
n⫺1
n 冊册
1
2
2
i
i
(i)
Finally, by substituting Equation (2.1) in Equation (i), the following expression for the variance results:
S2 ⫽
兺 y2i ⫺ ny2
n⫺1
(2.10)
Using Equation (2.10), the variance of a sample data set can be computed
by subtracting n times the square of the data’s mean from the summation of
the squared individual observations. With this equation, the variance and the
standard deviation can be computed directly from the data. However, it should
be stated that with large numerical values, Equation (2.10) may overwhelm
a handheld calculator or a computer working in single precision. If this problem should arise, the data should be centered or Equation (2.5) used. Cen-
2.9
NUMERICAL EXAMPLES
23
tering a data set involves subtracting a constant value (usually, the arithmetic
mean) from all values in a data set. By doing this, the values are modified to
a smaller, more manageable size.
2.9
NUMERICAL EXAMPLES
Example 2.1 Using the data from Table 2.2, determine the sample set’s
mean, median, and mode and the standard deviation using both Equations
(2.7) and (2.10). Also plot its histogram. (Recall that the data of Table 2.2
result from the seconds’ portion of 50 theodolite directions.)
SOLUTION
Mean: From Equation (2.1) and using the 兺 yi value from Table 2.4, we
have
y⫽
兺50
i⫽1 yi
50
⫽
1175
⫽ 23.5ⴖ
50
Median: Since there is an even number of observations, the data’s midpoint
lies between the values that are the 25th and 26th numerically from the beginning of the ordered set. These values are 23.4ⴖ and 23.5ⴖ, respectively.
Averaging these observations yields 23.45ⴖ.
Mode: The mode, which is the most frequently occurring value, is 23.8ⴖ.
It appears three times in the sample.
Range, class width, histogram: These data were developed in Section 2.4,
with the histogram plotted in Figure 2.1.
Standard deviation: Table 2.4 lists the residuals [computed using Equation
(2.3)] and their squares for each observation.
From Equation (2.7) and using the value of 92.36 from Table 2.4 as the
sum of the squared residuals, the standard deviation for the sample set is
computed as
S⫽
兺 v2i
⫽ 1.37ⴖ
冪n ⫺ 1 ⫽ 冪5092.36
⫺1
Summing the squared y-values of Table 2.4 yields
冘 y ⫽ 27,704.86
2
i
Using Equation (2.10), the standard deviation for the sample set is
24
TABLE 2.4 Data Arranged for the Solution of Example 2.1
No.
y
v
v2
No.
y
v
v2
No.
y
v
v2
No.
y
v
v2
1
2
3
4
5
6
7
8
9
10
11
12
20.1
20.5
21.2
21.7
21.8
21.9
22.0
22.2
22.3
22.3
22.5
22.6
3.4
3.0
2.3
1.8
1.7
1.6
1.5
1.3
1.2
1.2
1.0
0.9
11.56
9.00
5.29
3.24
2.89
2.56
2.25
1.69
1.44
1.44
1.00
0.81
13
14
15
16
17
18
19
20
21
22
23
24
22.6
22.7
22.8
22.8
22.9
22.9
23.0
23.1
23.1
23.2
23.2
23.3
0.9
0.8
0.7
0.7
0.6
0.6
0.5
0.4
0.4
0.3
0.3
0.2
0.81
0.64
0.49
0.49
0.36
0.36
0.25
0.16
0.16
0.09
0.09
0.04
25
26
27
28
29
30
31
32
33
34
35
36
37
23.4
23.5
23.6
23.7
23.8
23.8
23.8
23.9
24.0
24.1
24.1
24.2
24.3
0.1
0.0
⫺0.1
⫺0.2
⫺0.3
⫺0.3
⫺0.3
⫺0.4
⫺0.5
⫺0.6
⫺0.6
⫺0.7
⫺0.8
0.01
0.00
0.01
0.04
0.09
0.09
0.09
0.16
0.25
0.36
0.36
0.49
0.64
38
39
40
41
42
43
44
45
46
47
48
49
50
24.4
24.6
24.7
24.8
25.0
25.2
25.3
25.3
25.4
25.5
25.9
25.9
26.1
1175
⫺0.9
⫺1.1
⫺1.2
⫺1.3
⫺1.5
⫺1.7
⫺1.8
⫺1.8
⫺1.9
⫺2.0
⫺2.4
⫺2.4
⫺2.6
0.81
1.21
1.44
1.69
2.25
2.89
3.24
3.24
3.61
4.00
5.76
5.76
6.76
92.36
0.0
2.9
S⫽
兺 y2i ⫺ ny2
冪
n⫺1
⫽
NUMERICAL EXAMPLES
冪
27,704.86 ⫺ 50(23.5)2
⫽
50 ⫺ 1
25
⫽ 1.37ⴖ
冪92.36
49
By demonstration in Example 2.1, it can be seen that Equations (2.7) and
(2.10) will yield the same standard deviation for a sample set. Notice that the
number of observations within a single standard deviation of the mean, that
is, between (23.5ⴖ ⫺ 1.37ⴖ) and (23.5ⴖ ⫹ 1.37ⴖ), or between 22.13ⴖ and 24.87ⴖ,
is 34. This represents 34/50 ⫻ 100%, or 68%, of all observations in the
sample and matches the theory noted earlier. Also note that the algebraic sum
of residuals is zero, as was demonstrated by Equation (b).
The histogram shown in Figure 2.1 plots class relative frequencies versus
class values. Notice how the values tend to be grouped about the central point.
This is an example of a precise data set.
Example 2.2 The data set shown below also represents the seconds’ portion
of 50 theodolite observations of a direction. Compute the mean, median, and
mode, and use Equation (2.10) to determine the standard deviation. Also
construct a histogram. Compare the data of this example with those of Example 2.1.
34.2
33.6
35.2
30.1
38.4
34.0
30.2
34.1
37.7
36.4
37.9
33.0
33.5
35.9
35.9
32.4
39.3
32.2
32.8
36.3
35.3
32.6
34.1
35.6
33.7
39.2
35.1
33.4
34.9
32.6
36.7
34.8
36.4
33.7
36.1
34.8
36.7
30.0
35.3
34.4
33.7
34.1
37.8
38.7
33.6
32.6
34.7
34.7
36.8
31.8
SOLUTION Table 2.5, which arranges each observation and its square in
ascending order, is first prepared.
Mean: y ⫽ 兺 yi /n ⫽ 1737.0/50 ⫽ 34.74ⴖ
Median: The median is between the 25th and 26th values, which are both
34.7ⴖ. Thus, the median is 34.7ⴖ.
Mode: The data have three different values that occur with a frequency of
three. Thus, the modes for the data set are the three values 32.6ⴖ, 33.7ⴖ, and
34.1ⴖ.
Range: The range of the data is 39.3ⴖ ⫺ 30.0ⴖ ⫽ 9.3ⴖ.
Class width: For comparison purposes, the class width of 0.86 is taken
because it was used for the data in Table 2.2. Since it is desired that the
histogram be centered about the data’s mean value, the central interval is
determined by adding and subtracting one-half of the class width (0.43) to
the mean. Thus, the central interval is from (34.74ⴖ ⫺ 0.43ⴖ), or 34.31ⴖ, to
(34.74ⴖ ⫹ 0.43ⴖ), or 35.17ⴖ. To compute the remaining class intervals, the
26
TABLE 2.5 Data Arranged for the Solution of Example 2.2
No.
y
y2
No.
y
y2
No.
y
y2
No.
y
y2
1
2
3
4
5
6
7
8
9
10
11
12
30.0
30.1
30.2
31.8
32.2
32.4
32.6
32.6
32.6
32.8
33.0
33.4
900.00
906.01
312.04
1011.24
1036.84
1049.76
1062.76
1062.76
1062.76
1075.84
1089.00
1115.56
13
14
15
16
17
18
19
20
21
22
23
24
33.5
33.6
33.6
33.7
33.7
33.7
34.0
34.1
34.1
34.1
34.2
34.4
1122.25
1128.96
1128.96
1135.69
1135.69
1135.36
1156.00
1162.81
1162.81
1162.81
1169.64
1183.36
25
26
27
28
29
30
31
32
33
34
35
36
37
34.7
34.7
34.8
34.8
34.9
35.1
35.2
35.3
35.3
35.6
35.9
35.9
36.1
1204.09
1204.09
1211.04
1211.04
1218.01
1232.01
1239.04
1246.09
1246.09
1267.36
1288.81
1288.81
1303.21
38
39
40
41
42
43
44
45
46
47
48
49
50
36.3
36.4
36.4
36.7
36.7
36.8
37.7
37.8
37.9
38.4
38.7
39.2
39.3
1737.0
1317.69
1324.96
1324.96
1346.89
1346.89
1354.24
1421.29
1428.84
1436.41
1474.56
1497.69
1536.64
1544.49
60,584.48
2.9
NUMERICAL EXAMPLES
27
class width is subtracted, or added, to the bounds of the computed intervals
as necessary until all the data are contained within the bounds of the intervals.
Thus, the interval immediately preceding the central interval will be from
(34.31ⴖ ⫺ 0.86ⴖ), or 33.45ⴖ, to 34.31ⴖ, and the interval immediately following
the central interval will be from 35.17ⴖ to (35.17ⴖ ⫹ 0.86ⴖ), or 36.03ⴖ. In a
similar fashion, the remaining class intervals were determined and a class
frequency chart was constructed as shown in Table 2.6. Using this table, the
histogram of Figure 2.3 was constructed.
Variance: By Equation (2.10), using the sum of observations squared in
Table 2.5, the sample variance is
S2 ⫽
兺 y2i ⫺ ny2
n⫺1
⫽
60,584.48 ⫺ 50(34.74)2
⫽ 4.92
50 ⫺ 1
and the sample standard deviation is
S ⫽ 兹4.92 ⫽ 2.22ⴖ
The number of observations that actually fall within the bounds of the
mean S (i.e., between 34.74ⴖ 2.22ⴖ) is 30. This is 60% of all the observations, and closely approximates the theoretical value of 68.3%. These
bounds and the mean value are shown as dashed lines in Figure 2.3.
Comparison: The data set of Example 2.2 has a larger standard deviation
(2.22ⴖ) than that of Example 2.1 (1.37ⴖ). The range for the data of Example 2.2 (9.3ⴖ) is also larger than that of Example 2.1 (6.0ⴖ). Thus, the data
set of Example 2.2 is less precise than that of Example 2.1. A comparison of
TABLE 2.6 Frequency Table for Example 2.2
Class
29.15–30.01
30.01–30.87
30.87–31.73
31.73–32.59
32.59–33.45
33.45–34.31
34.31–35.17
35.17–36.03
36.03–36.89
36.89–37.75
37.75–38.61
38.61–39.47
Class
Frequency
Relative Class
Frequency
1
2
0
3
6
11
7
6
7
1
3
3
50
0.02
0.04
0.00
0.06
0.12
0.22
0.14
0.12
0.14
0.02
0.06
0.06
1.00
28
OBSERVATIONS AND THEIR ANALYSIS
Figure 2.3 Histogram for Example 2.2.
the two histograms shows this precision difference graphically. Note, for example, that the histogram in Figure 2.1 is narrower in width and taller at the
center than the histogram in Figure 2.3.
2.10 DERIVATION OF THE SAMPLE VARIANCE
(BESSEL’S CORRECTION)
Recall from Section 2.7 that the denominator of the equation for sample
variance was n ⫺ 1, whereas the denominator of the population variance was
n. A simple explanation for this difference is that one observation is necessary
to compute the mean (y), and thus only n ⫺ 1 observations remain for the
computation of the variance. A derivation of Equation (2.5) will clarify.
Consider a sample size of n drawn from a population with a mean, , and
standard error of . Let yi be an observation from the sample. Then
yi ⫺ ⫽ (yi ⫺ y) ⫹ (y ⫺ )
⫽ (yi ⫺ y) ⫹ ε
( j)
where ε ⫽ y ⫺ is the error or deviation of the sample mean. Squaring and
expanding Equation ( j) yields
(yi ⫺ )2 ⫽ (yi ⫺ y)2 ⫹ ε2 ⫹ 2ε(yi ⫺ y)
Summing all the observations in the sample from i equaling 1 to n yields
2.11
PROGRAMMING
冘 (y ⫺ ) ⫽ 冘 (y ⫺ y) ⫹ nε ⫹ 2ε 冘 (y ⫺ y)
n
n
n
2
2
i
2
i
i⫽1
29
i
i⫽1
(k)
i⫽1
Since by definition of the sample mean y
冘 (y ⫺ y) ⫽ 冘 y ⫺ ny ⫽ 冘 y ⫺ 冘 y ⫽ 0
n
n
n
i
n
i
i⫽1
i
i⫽1
i
i⫽1
(l)
i⫽1
Equation (k) becomes
冘 (y ⫺ ) ⫽ 冘 (y ⫺ y) ⫹ nε
n
n
2
2
i
2
i
i⫽1
(m)
i⫽1
Repeating this calculation for many samples, the mean value of the lefthand side of Equation (m) will (by definition of 2) tend to n2. Similarly,
by Equation (2.8), the mean value of nε2 ⫽ n( ⫺ y)2 will tend to n times
the variance of y since ε represents the deviation of the sample mean from
the population mean. Thus, nε2 → n(2 /n), where 2 /n is the variance in y
as n → ⬁. The discussion above, Equation (m) results in
冘 (y ⫺ y) ⫹
n
n2 →
2
2
(n)
2
(o)
i
i⫽1
Rearranging Equation (n) produces
冘 (y ⫺ y) → (n⫺ 1)
n
2
i
i⫽1
Thus, from Equation (o) and recognizing the left side of the equation as
(n ⫺ 1)S2 for a sample set of data, it follows that
S2 ⫽
兺ni⫽1 (yi ⫺ y)2
n⫺1
→ 2
(p)
In other words for a large number of random samples, the value of 兺ni⫽1
(yi ⫺ y)2 /(n ⫺ 1) tends to 2. That is, S2 is an unbiased estimate of the
population’s variance.
2.11
PROGRAMMING
STATS, a Windows-based statistical software package, is included on the CD
accompanying this book. It can be used to quickly perform statistical analysis
of data sets as presented in this chapter. Directions regarding its use are
provided on its help screen.
30
OBSERVATIONS AND THEIR ANALYSIS
Additionally, an electronic book is provided on the CD accompanying this
book. To view the electronic book interactively, Mathcad software is required.
However, for those of you who do not have a copy of Mathcad, html files of
the electronic book are included on the CD. The electronic book demonstrates
most of the numerical examples given in the book.
Many chapters include programming problems following the problem sets
at the end of the chapters. The electronic book demonstrates the rudiments
of programming these problems. Other programs on the CD include MATRIX
and ADJUST. MATRIX can be used to solve problems in the book that involve matrices. ADJUST has examples of working least squares adjustment
programs. ADJUST can be used to check solutions to many of the examples
in the book.
When you select the desired installation options, the installation program
provided on the CD will load the files to your computer. The installation
package will install each option as the option is selected. This software does
not remove (uninstall) the packages. This can be done using the ‘‘Add/Remove programs’’ options in your computer’s control panel.
PROBLEMS
2.1
The optical micrometer of a precise differential level is set and read
10 times as 8.801, 8.803, 8.798, 8.801, 8.799, 8.802, 8.802, 8.804,
8.800, and 8.802. What value would you assign to the operator’s ability
to set the micrometer on this instrument?
2.2
The seconds’ part of 50 pointings and readings for a particular direction
made using a 1ⴖ total station with a 0.1ⴖ display are
26.7, 26.4, 24.8, 27.4, 25.8, 27.0, 26.3, 27.8, 26.7, 26.0, 25.9, 25.4,
28.0, 27.2, 25.3, 27.2, 27.0, 27.7, 27.3, 24.8, 26.7, 25.3, 26.9, 25.5,
27.4, 25.4, 25.8, 25.5, 27.4, 27.2, 27.1, 27.4, 26.6, 26.2, 26.3, 25.3,
25.1, 27.3, 27.3, 28.1, 27.4, 27.2, 27.2, 26.4, 28.2, 25.5, 26.5, 25.9,
26.1, 26.3
(a) What is the mean of the data set?
(b) Construct a frequency histogram of the data using seven uniformwidth class intervals.
(c) What are the variance and standard deviation of the data?
(d) What is the standard deviation of the mean?
2.3
An EDM instrument and reflector are set at the ends of a baseline that
is 400.781 m long. Its length is measured 24 times, with the following
results:
PROBLEMS
31
400.787 400.796 400.792 400.787 400.787 400.786 400.792 400.794
400.790 400.788 400.797 400.794 400.789 400.785 400.791 400.791
400.793 400.791 400.792 400.787 400.788 400.790 400.798 400.789
(a) What are the mean, median, and standard deviation of the data?
(b) Construct a histogram of the data with five intervals and describe
its properties. On the histogram, lay off the sample standard deviation from both sides of the mean.
(c) How many observations are between y S, and what percentage
of observations does this represent?
2.4
Answer Problem 2.3 with the following additional observations:
400.784, 400.786, 400.789, 400.794, 400.792, and 400.789.
2.5
Answer Problem 2.4 with the following additional observations:
400.785, 400.793, 400.791, and 400.789.
2.6
A distance was measured in two parts with a 100-ft steel tape and then
in its entirety with a 200-ft steel tape. Five repetitions were made by
each method. What are the mean, variance, and standard deviation for
each method of measurement?
Distances measured with a 100-ft tape:
Section 1: 100.006, 100.004, 100.001, 100.006, 100.005
Section 2: 86.777, 86.779, 86.785, 86.778, 86.774
Distances measured with a 200-ft tape:
186.778, 186.776, 186.781, 186.766, 186.789
2.7
Repeat Problem 2.6 using the following additional data for the 200-ft
taped distance: 186.781, 186.794, 186.779, 186.778, and 186.776.
2.8
During a triangulation project, an observer made 16 readings for each
direction. The seconds’ portion of the directions to Station Orion are
listed as 43.0, 41.2, 45.0, 43.4, 42.4, 52.5, 53.6, 50.9, 52.0, 50.8, 51.9,
49.5, 51.6, 51.2, 51.8, and 50.2.
(a) Using a 1ⴖ class interval, plot the histogram using relative frequencies for the ordinates.
(b) Analyze the data and note any abnormalities.
(c) As a supervisor, would you recommend that the station be
reobserved?
2.9
Use the program STATS to compute the mean, median, mode, and
standard deviation of the data in Table 2.2 and plot a centered histogram of the data using nine intervals.
2.10 The particular line in a survey is measured three times on four separate
occasions. The resulting 12 observations in units of meters are 536.191,
32
OBSERVATIONS AND THEIR ANALYSIS
536.189, 536.187, 536.202, 536.200, 536.203, 536.202, 536.201,
536.199, 536.196, 536.205, and 536.202.
(a) Compute the mean, median, and mode of the data.
(b) Compute the variance and standard deviation of the data.
(c) Using a class width of 0.004 m, plot a histogram of the data and
note any abnormalities that may be present.
Use the program STATS to do:
2.11 Problem 2.2.
2.12 Problem 2.3.
2.13 Problem 2.4.
2.14 Problem 2.5.
2.15 Problem 2.10. Use a class width of 0.003 m in part (c).
Practical Exercises
2.16 Using a total station, point and read a horizontal circle to a well-defined
target. With the tangent screw or jog-shuttle mechanism, move the
instrument of the point and repoint on the same target. Record this
reading. Repeat this process 50 times. Perform the calculations of Problem 2.2 using this data set.
2.17 Determine your EDM–reflector constant, K, by observing the distances
between the following three points:
A
B
C
Distance AC should be roughly 1 mile long, with B situated at some
location between A and C. From measured values AC, AB, and BC, the
constant K can be determined as follows: Since
AC ⫹ K ⫽ (AB ⫹ K) ⫹ (BC ⫹ K)
thus
K ⫽ AC ⫺ (AB ⫹ BC)
When establishing the line, be sure that AB ⫽ BC and that all three
points are precisely on a straight line. Use three tripods and tribrachs
to minimize setup errors and be sure that all are in adjustment. Measure
each line 20 times with the instrument in the metric mode. Be sure to
adjust the distances for the appropriate temperature and pressure and
for differences in elevation. Determine the 20 values of K and analyze
the sample set. What is the mean value for K and its standard
deviation?
CHAPTER 3
RANDOM ERROR THEORY
3.1
INTRODUCTION
As noted earlier, the adjustment of measured quantities containing random
errors is a major concern to people involved in the geospatial sciences. In the
remaining chapters it is assumed that all systematic errors have been removed
from the measured values and that only random errors and blunders (which
have escaped detection) remain. In this chapter the general theory of random
errors is developed, and some simple methods that can be used to isolate
remaining blunders in sets of data are discussed.
3.2
THEORY OF PROBABILITY
Probability is the ratio of the number of times that an event should occur to
the total number of possibilities. For example, the probability of tossing a
two with a fair die is 1/6 since there are six total possibilities (faces on a
die) and only one of these is a two. When an event can occur in m ways and
fail to occur in n ways, the probability of its occurrence is m/(m ⫹ n), and
the probability of its failure is n/(m ⫹ n).
Probability is always a fraction ranging between zero and one. Zero denotes impossibility, and one indicates certainty. Since an event must either
occur or fail to occur, the sum of all probabilities for any event is 1, and thus
if 1/6 is the probability of throwing a two with one throw of a die, then 1 ⫺
1/6, or 5/6, is the probability that a two will not appear.
In probability terminology, a compound event is the simultaneous occurrence of two or more independent events. This is the situation encountered
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
33
34
RANDOM ERROR THEORY
most frequently in surveying. For example, random errors from angles and
distances (compound events) cause traverse misclosures. The probability of
the simultaneous occurrence of two independent events is the product of their
individual probabilities.
To illustrate this condition, consider the simple example of having two
boxes containing combinations of red and white balls. Box A contains four
balls, one red and three white. Box B contains five balls, two red and three
white. What is the probability that two red balls would be drawn if one ball
is drawn randomly from each box? The total number of possible pairs is 4
⫻ 5 (20) since by drawing one ball from box A, any of the five balls in box
B would complete the pair. There are only two ways to draw two red balls;
that is, box A’s red ball can be matched with either red ball from box B.
Therefore, the probability of obtaining two red balls simultaneously is 2/20.
Thus, the probability of this compound event is the product of the individual
probabilities of drawing a red ball from each box, or
P ⫽ 1/4 ⫻ 2/5 ⫽ 2/20
Similarly, the probability of drawing two white balls simultaneously is 3/4
⫻ 3/5, or 9/20, and the probability of getting one red ball and one white
ball is 1 ⫺ (2/20 ⫹ 9/20), or 9/20.
From the foregoing it is seen that the probability of the simultaneous occurrence of two independent events is the product of the individual probabilities of those two events. This principle is extended to include any number
of events:
P ⫽ P1 ⫻ P2 ⫻ 䡠 䡠 䡠 ⫻ Pn
(3.1)
where P is the probability of the simultaneous occurrence of events having
individual probabilities P1, P2, . . . , Pn.
To develop the principle of how random errors occur, consider a very
simple example where a single tape measurement is taken between points A
and B. Assume that this measurement contains a single random error of size
1. Since the error is random, there are two possibilities for its value ⫹1 or
⫺1. Let t be the number of ways that each error can occur, and let T be the
total number of possibilities, which is two. The probability of obtaining ⫹1,
which can occur only one way (i.e., t ⫽ 1), is t/T or 1/2. This is also the
probability of obtaining ⫺1. Suppose now that in measuring a distance AE,
the tape must be placed end to end so that the result depends on the combination of two of these tape measurements. Then the possible error combinations in the result are ⫺1 and ⫺1, ⫺1 and ⫹1, ⫹1 and ⫺1, and ⫹1 and ⫹1,
with T ⫽ 4. The final errors are ⫺2, 0, and ⫹2, and their t values are 1, 2,
and 1, respectively. This produces probabilities of 1/4, 1/2, and 1/4, respectively. In general, as n, the number of single combined measurements, is
increased, T increases according to the function T ⫽ 2n, and thus for three
3.2
THEORY OF PROBABILITY
35
combined measurements, T ⫽ 23 ⫽ 8, and for four measurements, T ⫽ 24 ⫽
16.
The analysis of the preceding paragraph can be continued to obtain the
results shown in Table 3.1. Figure 3.1(a) through (e) are histogram plots of
the results in Table 3.1, where the values of the errors are plotted as the
abscissas and the probabilities are plotted as ordinates of equal-width bars.
If the number of combining measurements, n, is increased progressively to
larger values, the plot of error sizes versus probabilities would approach a
smooth curve of the characteristic bell shape shown in Figure 3.2. This curve
is known as the normal error distribution curve. It is also called the probability density function of a normal random variable. Notice that when n is 4,
as illustrated in Figure 3.1(d), and when n ⫽ 5, as shown in Figure 3.1(e),
the dashed lines are already beginning to take on this form.
It is important to notice that the total area of the vertical bars for each plot
equals 1. This is true no matter the value of n, and thus the area under the
smooth normal error distribution curve is equal to 1. If an event has a probability of 1, it is certain to occur, and therefore the area under the curve
represents the sum of all the probabilities of the occurrence of errors.
TABLE 3.1 Occurrence of Random Errors
(1)
Number of
Combining
Measurements
(2)
Value of
Resulting
Error
1
⫹1
⫺1
⫹2
2
3
4
5
0
⫺2
⫹3
⫹1
⫺1
⫺3
⫹4
⫹2
0
⫺2
⫺4
⫹5
⫹3
⫹1
⫺1
⫺3
⫺5
(3)
Frequency,
t
1
1
1
2
1
1
3
3
1
1
4
6
4
1
1
5
10
10
5
1
(4)
Total
Number of
Possibilities,
T
2
4
8
16
32
(5)
Probability
1/2
1/2
1/4
1/2
1/4
1/8
3/8
3/8
1/8
1 / 16
1/4
3/8
1/4
1 / 16
1 / 32
5 / 32
5 / 16
5 / 16
5 / 32
1 / 32
36
RANDOM ERROR THEORY
Figure 3.1 Plots of probability versus size of errors.
As derived in Section D.1, the equation of the normal distribution curve,
also called the normal probability density function, is
ƒ(x) ⫽
1
2
2
e⫺x / 2
兹2
(3.2)
where ƒ(x) is the probability density function, e the base of natural logarithms,
x the error, and the standard error as defined in Chapter 2.
3.3
PROPERTIES OF THE NORMAL DISTRIBUTION CURVE
In Equation (3.2), ƒ(x) is the probability of occurrence of an error of size
between x and x ⫹ dx, where dx is an infinitesimally small value. The error’s
probability is equivalent to the area under the curve between the limits of x
and x ⫹ dx, which is shown crosshatched in Figure 3.3. As stated previously,
the total area under the probability curve represents the total probability,
which is 1. This is represented in equation form as
Figure 3.2 The normal distribution curve.
3.3
PROPERTIES OF THE NORMAL DISTRIBUTION CURVE
37
Figure 3.3 Normal density function.
area ⫽
冕
⬁
1
⫺⬁
兹2
2 / 22
e⫺x
dx ⫽ 1
(3.3)
Let y represent ƒ(x) in Equation (3.2) and differentiate:
冉
dy
x
1
2
2
⫽⫺ 2
ex / 2
dx
兹2
冊
(3.4)
Recognizing the term in parentheses in Equation (3.4) as y gives
dy
x
⫽⫺ 2y
dx
(3.5)
Taking the second derivative of Equation (3.2), we obtain
d 2y
x dy
y
⫽⫺ 2
⫺ 2
dx2
dx
(3.6)
Substituting Equation (3.5) into Equation (3.6) yields
d 2y x2
y
⫽ 4y⫺ 2
2
dx
(3.7)
Equation (3.7) can be simplified to
冉
冊
d 2y
y x2
⫽
⫺1
dx2 2 2
(3.8)
38
RANDOM ERROR THEORY
From calculus, the first derivative of a function yields the slope of the
function when evaluated at a point. In Equation (3.5), dy/dx ⫽ 0 when the
values of x or y equal zero. This implies that the curve is parallel to the x
axis at the center of the curve when x is zero and is asymptotic to the x axis
as y approaches zero.
Also from calculus, a function’s second derivative provides the rate of
change in a slope when evaluated at a point. The curve’s inflection points
(points where the algebraic sign of the slope changes) can be located by
finding where the function’s second derivative equals zero. In Equation (3.8),
d2y/dx2 ⫽ 0 when x2 / 2 ⫺ 1 ⫽ 0, and thus the curve’s inflection point occurs
when x equals Ⳳ.
Since e0 ⫽ 1, if x is set equal to zero, Equation (3.2) gives us
y⫽
1
兹2
(3.9)
This is the curve’s central ordinate, and as can be seen, it is inversely proportional to . According to Equation (3.9), a group of measurements having
small must have a large central ordinate. Thus, the area under the curve
will be concentrated near the central ordinate, and the errors will be correspondingly small. This indicates that the set of measurements is precise. Since
bears this inverse relationship to the precision, it is a numerical measure
for the precision of a measurement set. In Section 2.7 we defined as the
standard error and gave equations for computing its value.
3.4
STANDARD NORMAL DISTRIBUTION FUNCTION
In Section 3.2 we defined the probability density function of a normal random
2
2
variable as ƒ(x) ⫽ 1/ (兹2)e⫺x / 2 . From this we develop the normal distribution function
冕
t
Fx(t) ⫽
⫺⬁
1
2
2
e⫺x / 2 dx
兹2
(3.10)
where t is the upper bound of integration, as shown in Figure 3.4.
As stated in Section 3.3, the area under the normal density curve represents
the probability of occurrence. Furthermore, integration of this function yields
the area under the curve. Unfortunately, the integration called for in Equation
(3.10) cannot be carried out in closed form, and thus numerical integration
techniques must be used to tabulate values for this function. This has been
done for the function when the mean is zero ( ⫽ 0) and the variance is 1
(2 ⫽ 1). The results of this integration are shown in the standard normal
distribution table, Table D.1. In this table the leftmost column with a heading
3.4
39
STANDARD NORMAL DISTRIBUTION FUNCTION
Figure 3.4 Area under the distribution curve determined by Equation (3.10).
of t is the value shown in Figure 3.4 in units of . The top row (with headings
0 through 9) represents the hundredths decimal places for the t values. The
values in the body of Table D.1 represent areas under the standard normal
distribution curve from ⫺⬁ to t. For example, to determine the area under
the curve from ⫺⬁ to 1.68, first find the row with 1.6 in the t column. Then
scan along the row to the column with a heading of 8. At the intersection of
row 1.6 and column 8 (1.68), the value 0.95352 occurs. This is the area under
the standard normal distribution curve from ⫺⬁ to a t value of 1.68. Similarly,
other areas under the standard normal distribution curve can be found for
various values for t. Since the area under the curve represents probability and
its maximum area is 1, this means that there is a 95.352% (0.95352 ⫻ 100%)
probability that t is less than or equal to 1.68. Alternatively, it can be stated
that there is a 4.648% [(1 ⫻ 0.95352) ⫻ 100%] probability that t is greater
than 1.68.
Once available, Table D.1 can be used to evaluate the distribution function
for any mean, , and variance, 2. For example, if y is a normal random
variable with a mean of and a variance of 2, an equivalent normal random
variable z ⫽ (y ⫺ )/ can be defined that has a mean of zero and a variance
of 1. Substituting the definition for z with ⫽ 0 and 2 ⫽ 1 into Equation
(3.2), its density function is
Nz(z) ⫽
1
2
e⫺z / 2
兹2
(3.11)
and its distribution function, known as the standard normal distribution function, becomes
冕
t
Nz(z) ⫽
⫺⬁
1
兹2
2/2
e⫺z
dz
(3.12)
For any group of normally distributed measurements, the probability of the
normal random variable can be computed by analyzing the integration of the
40
RANDOM ERROR THEORY
distribution function. Again, as stated previously, the area under the curve in
Figure 3.4 represents probability. Let z be a normal random variable, then the
probability that z is less than some value of t is given by
P(z ⬍ t) ⫽ Nz(t)
(3.13)
To determine the area (probability) between t values of a and b (the crosshatched area in Figure 3.5), the difference in the areas between a and b,
respectively, can be computed. By Equation (3.13), the area from ⫺⬁ to b is
P(z ⬍ b) ⫽ Nz(b). By the same equation, the area from ⫺⬁ to a is P(z ⬍ a)
⫽ Nz(a). Thus, the area between a and b is the difference in these values and
is expressed as
P(a ⬍ z ⬍ b) ⫽ Nz(b) ⫺ Nz(a)
(3.14)
If the bounds are equal in magnitude but opposite in sign (i.e., ⫺a ⫽ b ⫽ t),
the probability is
P(兩z兩 ⬍ t) ⫽ Nz(t) ⫺ Nz(⫺t)
(3.15)
From the symmetry of the normal distribution in Figure 3.6 it is seen that
P(z ⬎ t) ⫽ P(z ⬍ ⫺t)
(3.16)
for any t ⬎ 0. This symmetry can also be shown with Table D.1. The tabular
value (area) for a t value of ⫺1.00 is 0.15866. Furthermore, the tabular value
for a t value of ⫹1.00 is 0.84134. Since the maximum probability (area) is
1, the area above ⫹1.00 is 1 ⫺ 0.84134, or 0.15866, which is the same as
the area below ⫺1.00. Thus, since the total probability is always 1, we can
define the following relationship:
1 ⫺ Nz(t) ⫽ Nz(⫺t)
Now substituting Equation (3.17) into Equation (3.15), we have
Figure 3.5 Area representing the probability in Equation (3.14).
(3.17)
3.5
PROBABILITY OF THE STANDARD ERROR
41
Figure 3.6 Area representing the probability in Equation (3.16).
P(兩z兩 ⬍ t) ⫽ 2Nz(t) ⫺ 1
3.5
(3.18)
PROBABILITY OF THE STANDARD ERROR
The equations above can be used to determine the probability of the standard
error, which from previous discussion is the area under the normal distribution
curve between the limits of Ⳳ. For the standard normal distribution when
2 is 1, it is necessary to locate the values of t ⫽ ⫺1 ( ⫽ ⫺1) and t ⫽ ⫹1
( ⫽ 1) in Table D.1. As seen previously, the appropriate value from the table
for t ⫽ ⫺1.00 is 0.15866. Also, the tabular value for t ⫽ 1.00 is 0.84134,
and thus, according to Equation (3.15), the area between ⫺ and ⫹ is
P(⫺ ⬍ z ⬍ ⫹) ⫽ Nz(⫹) ⫺ Nz(⫺)
⫽ 0.84134 ⫺ 0.15866 ⫽ 0.68268
From this it has been determined that about 68.3% of all measurements
from any data set are expected to lie between ⫺ and ⫹. It also means that
for any group of measurements there is approximately a 68.3% chance that
any single observation has an error between plus and minus . The crosshatched area of Figure 3.7 illustrates that approximately 68.3% of the area
Figure 3.7 Normal distribution curve.
42
RANDOM ERROR THEORY
exists between plus and minus . This is true for any set of measurements
having normally distributed errors. Note that as discussed in Section 3.3, the
inflection points of the normal distribution curve occur at Ⳳ. This is illustrated in Figure 3.7.
3.5.1
50% Probable Error
For any group of observations, the 50% probable error establishes the limits
within which 50% of the errors should fall. In other words, any measurement
has the same chance of coming within these limits as it has of falling outside
them. Its value can be obtained by multiplying the standard deviation of the
observations by the appropriate t value. Since the 50% probable error has a
probability of 1/2, Equation (3.18) is set equal to 0.50 and the t value corresponding to this area is determined as
P(兩z兩 ⬍ t) ⫽ 0.5 ⫽ 2Nz(t) ⫺ 1
1.5 ⫽ 2Nz(t)
0.75 ⫽ Nz(t)
From Table D.1 it is apparent that 0.75 is between a t value of 0.67 and 0.68;
that is,
Nz(0.67) ⫽ 0.7486
and Nz(0.68) ⫽ 0.7517
The t value can be found by linear interpolation, as follows:
t
0.75 ⫺ 0.7486
0.0014
⫽
⫽
⫽ 0.4516
0.68 ⫺ 0.67 0.7517 ⫺ 0.7486 0.0031
t ⫽ 0.01 ⫻ 0.4516
and t ⫽ 0.67 ⫹ 0.0045 ⫽ 0.6745.
For any set of observations, therefore, the 50% probable error can be obtained by computing the standard error and then multiplying it by 0.6745:
E50 ⫽ 0.6745
3.5.2
(3.19)
95% Probable Error
The 95% probable error, E95, is the bound within which, theoretically, 95%
of the observation group’s errors should fall. This error category is popular
with surveyors for expressing precision and checking for outliers in data.
3.6
43
USES FOR PERCENT ERRORS
Using the same reasoning as in developing the equation for the 50% probable
error, substituting into Equation (3.18) gives
0.95 ⫽ P(兩z兩 ⬍ t) ⫽ 2Nz(t) ⫺ 1
1.95 ⫽ 2Nz(t)
0.975 ⫽ Nz(t)
Again from the Table D.1, it is determined that 0.975 occurs with a t value
of 1.960. Thus, to find the 95% probable error for any group of measurements,
the following equation is used:
E95 ⫽ 1.960
3.5.3
(3.20)
Other Percent Probable Errors
Using the same computational techniques as in Sections 3.5.1 and 3.5.2, other
percent probable errors can be calculated. One other percent error worthy of
particular note is E99.7. It is obtained by multiplying the standard error by
2.968:
E99.7 ⫽ 2.968
(3.21)
This value is often used for detecting blunders, as discussed in Section 3.6.
A summary of probable errors with varying percentages, together with their
multipliers, is given in Table 3.2.
3.6
USES FOR PERCENT ERRORS
Standard errors and errors of other percent probabilities are commonly used
to evaluate measurements for acceptance. Project specifications and contracts
TABLE 3.2 Multipliers for Various Percent Probable Errors
Symbol
Multiplier
Percent Probable
Errors
E50
E90
E95
E99
E99.7
E99.9
0.6745
1.645
1.960
2.576
2.968
3.29
50
90
95
99
99.7
99.9
44
RANDOM ERROR THEORY
often require that acceptable errors be within specified limits, such as the 90%
and 95% errors. The 95% error, sometimes called the two-sigma (2) error
because it is computed as approximately 2, is most often specified. Standard
error is also frequently used. The probable error, E50, is seldom employed.
Higher percent errors are used to help isolate outliers (very large errors)
and blunders in data sets. Since outliers seldom occur in a data set, measurements outside a selected high percentage range can be rejected as possible
blunders. Generally, any data that differ from the mean by more than 3 can
be considered as blunders and removed from a data set. As seen in Table 3.2,
rejecting observations greater that 3 means that about 99.7% of all measurements should be retained. In other words, only about 0.3% of the measurements in a set of normally distributed random errors (or 3 observations
in 1000) should lie outside the range Ⳳ3.
Note that as explained in Chapter 2, standard error and standard deviation
are often used interchangeably, when in practice it is actually the standard
deviation that is computed, not the standard error. Thus, for practical applications, in the equations of the preceding sections is replaced by S to
distinguish between these two related values.
3.7
PRACTICAL EXAMPLES
Example 3.1 Suppose that the following values (in feet) were obtained in
15 independent distance observations, Di: 212.22, 212.25, 212.23, 212.15,
212.23, 212.11, 212.29, 212.34, 212.22, 212.24, 212.19, 212.25, 212.27,
212.20, and 212.25. Calculate the mean, S, E50, E95, and check for any observations outside the 99.7% probability level.
SOLUTION From Equation (2.1), the mean is
D⫽
兺 Di
n
⫽
3183.34
⫽ 212.22 ft
15
From Equation (2.10), S is
S⫽
冪
675,576.955 ⫺ 15(212.2232)
⫽
15 ⫺ 1
⫽ Ⳳ0.055 ft
冪0.051298
14
where 兺 Di ⫽ 675,576.955. By scanning the data, it is seen that 10 observations are between 212.22 Ⳳ 0.06 ft or within the range (212.16, 212.28).1
1
The expression (x, y) represents a range between x and y. That is, the population mean lies
between 212.16 and 212.28 in this example.
3.7
PRACTICAL EXAMPLES
45
This corresponds to 10/15 ⫻ 100, or 66.7% of the observations. For the set,
this is what is expected if it conforms to normal error distribution theory.
From Equation (3.19), E50 is
E50 ⫽ 0.6745S ⫽ Ⳳ0.6745 (0.055) ⫽ Ⳳ0.04 ft
Again by scanning the data, nine observations lie in the range 212.22 Ⳳ
0.04 ft. That is, they are within the range (212.18, 212.26) ft. This corresponds
to 9/15 ⫻ 100%, or 60% of the observations. Although this should be 50%
and thus is a little high for a normal distribution, it must be remembered that
this is only a sample of the population and should not be considered a reason
to reject the entire data set. (In Chapter 4, statistical intervals involving sample
sets are discussed.)
From Equation (3.20), E95 is
E95 ⫽ 1.960S ⫽ Ⳳ1.960(0.055) ⫽ Ⳳ0.11 ft
Note that 14 of the observations lie in the range 212.22 Ⳳ 0.11 (212.11,
212.33) ft, or 93% of the data is in the range.
At the 99.7% level of confidence, the range Ⳳ2.968S corresponds to an
interval of Ⳳ0.16 ft. With this criterion for rejection of outliers, all values in
the data lie in this range. Thus, there is no reason to believe that any observation is a blunder or outlier.
Example 3.2 The seconds portion of 50 micrometer readings from a 1ⴖ
theodolite are listed below. Find the mean, standard deviation, and E95. Check
the observations at a 99% level of certainty for blunders.
41.9 46.3 44.6 46.1 42.5 45.9 45.0 42.0 47.5 43.2 43.0 45.7 47.6
49.5 45.5 43.3 42.6 44.3 46.1 45.6 52.0 45.5 43.4 42.2 44.3 44.1
42.6 47.2 47.4 44.7 44.2 46.3 49.5 46.0 44.3 42.8 47.1 44.7 45.6
45.5 43.4 45.5 43.1 46.1 43.6 41.8 44.7 46.2 43.2 46.8
SOLUTION The sum of the 50 observations is 2252, and thus the mean is
2252/50 ⫽ 45.04ⴖ. Using Equation (2.10), the standard deviation is
S⫽
冪
101,649.94 ⫺ 50(45.04)2
⫽ Ⳳ2.12ⴖ
50 ⫺ 1
46
RANDOM ERROR THEORY
where 兺 y2 ⫽ 101,649.94. There are 35 observations in the range 45.04ⴖ Ⳳ
2.12ⴖ, or from 42.92ⴖ to 47.16ⴖ. This corresponds to 35/50 ⫻ 100 ⫽ 70% of
the observations and correlates well with the anticipated level of 68.3%.
From Equation (3.20), E95 ⫽ Ⳳ1.960(2.12ⴖ) ⫽ Ⳳ4.16ⴖ. The data actually
contain three values that deviate from the mean by more than 4.16ⴖ (i.e., that
are outside the range 40.88ⴖ to 49.20ⴖ). They are 49.5ⴖ (two values) and 52.0ⴖ.
No values are less than 40.88ⴖ, and therefore 47/50 ⫻ 100%, or 94% of the
observations lie in the E95 range.
From Equation (3.21), E99 ⫽ Ⳳ2.576(2.12ⴖ) ⫽ Ⳳ5.46ⴖ, and thus 99% of
the data should fall in the range 45.04ⴖ Ⳳ 5.46ⴖ, or (39.58ⴖ, 50.50ⴖ). Actually,
one value is greater than 50.50ⴖ, and thus 98% of all the observations fall in
this range.
By the analysis above it is seen that the data set is skewed to the left. That
is, values higher than the range always fell on the right side of the data. The
histogram shown in Figure 3.8 depicts this skewness. This suggests that it
may be wise to reject the value of 52.0ⴖ as a mistake. The recomputed values
for the data set (minus 52.0ⴖ) are
mean ⫽
2252ⴖ ⫺ 52ⴖ
⫽ 44.90ⴖ
49
兺 y2 ⫽ 101,649.94 ⫺ 52.02 ⫽ 98,945.94
S⫽
冪
98,945.94 ⫺ 49(44.89795918)2
⫽ Ⳳ1.88ⴖ
49 ⫺ 1
Now after recomputing errors, 32 observations lie between plus S or minus
S, which represents 65.3% of the observations, 47 observations lie in the E95
Figure 3.8 Skewed data set.
PROBLEMS
47
range, which represents 95.9% of the data, and no values are outside the E99
range. Thus, there is no reason to reject any additional data at a 99% level
of confidence.
PROBLEMS
3.1
Determine the t value for E80.
3.2
Determine the t value for E75.
3.3
Use STATS to determine t for E90.
3.4
Use STATS to determine t for E99.9.
3.5
Assuming a normal distribution, explain the statement: ‘‘As the standard deviation of the group of observations decreases, the precision of
the group increases.’’
3.6
If the mean of a population is 2.456 and its variance is 2.042, what is
the peak value for the normal distribution curve and the points of
inflection?
3.7
If the mean of a population is 13.4 and its variance is 5.8, what is the
peak value for the normal distribution curve and the points of
inflection?
3.8
Plot the curve in Problem 3.6 using Equation (3.2) to determine ordinate and abscissa values.
3.9
Plot the curve in Problem 3.7 using Equation (3.2) to determine ordinate and abscissa values.
3.10 The following data represent 60 planimeter observations of the area
within a plotted traverse.
1.677 1.676 1.657 1.667 1.673 1.671 1.673 1.670 1.675 1.664 1.664 1.668
1.664 1.651 1.663 1.665 1.670 1.671 1.651 1.665 1.667 1.662 1.660 1.667
1.660 1.667 1.667 1.652 1.664 1.690 1.649 1.671 1.675 1.653 1.654 1.665
1.668 1.658 1.657 1.690 1.666 1.671 1.664 1.685 1.667 1.655 1.679 1.682
1.662 1.672 1.667 1.667 1.663 1.670 1.667 1.669 1.671 1.660 1.683 1.663
(a) Calculate the mean and standard deviation.
(b) Plot the relative frequency histogram (of residuals) for the data
above using a class interval of one-half of the standard deviation.
(c) Calculate the E50 and E90 intervals.
48
RANDOM ERROR THEORY
(d) Can any observations be rejected at a 99% level of certainty?
(e) What is the peak value for the normal distribution curve, and where
are the points of inflection on the curve?
3.11 Discuss the normality of each set of data below and whether any observations may be removed at the 99% level of certainty as blunders
or outliers. Determine which set is more precise after apparent blunders
and outliers are removed. Plot the relative frequency histogram to defend your decisions.
Set 1:
468.09 468.13 468.11 468.13 468.10 468.13 468.12 468.09 468.14
468.10 468.10 468.12 468.14 468.16 468.12 468.10 468.10 468.11
468.13 468.12 468.18
Set 2:
750.82 750.86 750.83 750.88 750.88 750.86 750.86 750.85 750.86
750.86 750.88 750.84 750.84 750.88 750.86 750.87 750.86 750.83
750.90 750.84 750.86
3.12 Using the following data set, answer the questions that follow.
17.5
15.0
13.4
23.9
25.2
19.5
25.8
30.0
22.5
35.3
39.5
23.5
26.5
21.3
22.3
21.6
27.2
21.1
24.0
23.5
32.5
32.2
24.2
35.7
28.0
24.0
16.8
21.1
19.0
30.7
30.2
33.7
19.7
19.7
25.1
27.9
28.5
22.7
31.0
28.4
31.2
24.6
30.2
16.8
26.9
23.3
21.5
18.8
21.4
20.7
(a) What are the mean and standard deviation of the data set?
(b) Construct a centered relative frequency histogram of the data using
seven intervals and discuss whether it appears to be a normal data
set.
(c) What is the E95 interval for this data set?
(d) Would there be any reason to question the validity of any observation at the 95% level?
PROBLEMS
49
3.13 Repeat Problem 3.12 using the following data:
2.898 2.918 2.907 2.889 2.901 2.901 2.899 2.899 2.911 2.909
2.904 2.905 2.895 2.920 2.899 2.896 2.907 2.897 2.900 2.897
Use STATS to do each problem.
3.14 Problem 3.10
3.15 Problem 3.11
3.16 Problem 3.12
3.17 Problem 3.13
Programming Problems
3.18 Create a computational package that solves Problem 3.11.
3.19 Create a computational package that solves Problem 3.12.
CHAPTER 4
CONFIDENCE INTERVALS
4.1
INTRODUCTION
Table 4.1 contains a discrete population of 100 values. The mean () and
variance (2) of that population are 26.1 and 17.5, respectively. By randomly
selecting 10 values from Table 4.1, an estimate of the mean and variance of
the population can be determined. However, it should not be expected that
these estimates (y and S2) will exactly match the mean and variance of the
population. Sample sets of 10 values each could continue to be selected from
the population to determine additional estimates for the mean and variance
of the population. However, it is just as unlikely that these additional values
would match those obtained from either the population or the first sample set.
As the sample size is increased, the mean and variance of the sample
should approach the values of the population. In fact, as the sample size
becomes very large, the mean and variance of the samples should be close
to those of the population. This procedure was carried out for various sample
sizes starting at 10 values and increasing the sample by 10 values, with the
results shown in Table 4.2. Note that the value computed for the mean of the
sample approaches the value of the population as the sample size is increased.
Similarly, the value computed for the variance of the sample tends to approach
the value of the population as the sample size is increased.
Since the mean of a sample set y and its variance S2 are computed from
random variables, they are also random variables. This means that even if the
size of the sample is kept constant, varying values for the mean and variance
can be expected from the samples, with greater confidence given to larger
samples. Also, it can be concluded that the values computed from a sample
also contain errors. To illustrate this, an experiment was run for four randomly
50
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
4.1
INTRODUCTION
51
TABLE 4.1 Population of 100 Values
18.2
25.7
30.0
30.7
24.3
29.0
28.0
26.5
27.1
26.8
22.0
21.3
30.3
30.2
21.3
26.4
25.2
26.5
32.2
24.4
21.9
24.0
31.5
27.0
27.7
18.4
28.8
26.5
18.9
26.7
20.1
26.3
28.1
22.2
29.0
25.4
19.4
28.0
25.2
39.8
26.4
22.8
26.9
28.9
29.9
26.7
25.6
29.2
25.0
27.3
27.0
22.4
24.0
19.8
24.2
28.5
26.6
27.6
29.8
30.6
20.3
26.1
29.9
23.4
32.0
23.4
24.5
29.3
29.9
30.9
28.2
19.6
26.6
22.6
35.5
26.8
25.2
38.2
27.3
21.2
23.8
28.5
21.8
19.1
24.2
27.9
26.2
22.3
22.9
25.3
20.8
22.6
15.3
27.7
28.2
24.7
36.0
28.1
25.5
24.9
selected sets of 10 values from Table 4.1. Table 4.3 lists the samples, their
means, and variances. Notice the variation in the values computed for the
four sets. As discussed above, this variation is expected.
Fluctuations in the means and variances computed from varying sample
sets raises questions about the ability of these values to estimate the population values reliably. For example, a higher confidence is likely to be placed
on a sample set with a small variance than on one with a large variance. Thus,
in Table 4.3, because of its small variance, one is more likely to believe that
the mean of the second sample set is a more reliable estimate than the others
for the mean of the population. In reality, this is not the case, since the means
of the other three sets are actually closer to the population mean of 26.1.
As noted earlier, the size of the sample should also be considered when
determining the reliability of a computed mean or variance. If the mean were
TABLE 4.2 Increasing Sample Sizes
No.
y
S2
10
20
30
40
50
60
70
80
90
100
26.9
25.9
25.9
26.5
26.6
26.4
26.3
26.3
26.3
26.1
28.1
21.9
20.0
18.6
20.0
17.6
17.1
18.4
17.8
17.5
52
CONFIDENCE INTERVALS
TABLE 4.3 Random Sample Sets from a Population
Set
Set
Set
Set
1:
2:
3:
4:
29.9,
26.9,
32.2,
24.2,
18.2,
28.1,
22.2,
36.0,
30.7,
29.2,
23.4,
18.2,
24.4,
26.2,
27.9,
24.3,
36.0,
30.0,
27.0,
24.0,
25.6,
27.1,
28.9,
28.9,
26.5,
26.5,
22.6,
28.8,
29.9,
30.6,
27.7,
30.2,
19.6,
28.5,
30.6,
28.1,
27.9
25.5
26.9
29.0
y
y
y
y
⫽
⫽
⫽
⫽
26.9,
27.9,
26.9,
27.2,
S2
S2
S2
S2
⫽
⫽
⫽
⫽
28.1
2.9
10.9
23.0
computed from a sample of five values, and another computed from a sample
of 30 values, more confidence is likely to be placed on the values derived
from the larger sample set than on those from the smaller one, even if both
sample sets have the same mean and standard deviation.
In statistics, this relationship between the sample sets, the number of samples, and the values computed for the means and variances is part of sampling
distribution theory. This theory recognizes that estimates for the mean and
variance do vary from sample to sample. Estimators are the functions used
to compute these estimates. Examples of estimator functions are Equations
(2.1) and (2.5), which are used to compute estimates of the mean and variance
for a population, respectively. As demonstrated and discussed, these estimates
vary from sample to sample and thus have their own population distributions.
In Section 4.2, three distributions are defined that are used for describing or
quantifying the reliability of mean and variance estimates. By applying these
distributions, statements can be written for the reliability at any given level
of confidence of the estimates computed. In other words, a range called the
confidence interval can be determined within which the population mean and
population variance can be expected to fall for varying levels of probability.
4.2
DISTRIBUTIONS USED IN SAMPLING THEORY
4.2.1 2 Distribution
The chi-square distribution, symbolized as 2, compares the relationship between the population variance and the variance of a sample set based on the
number of redundancies, , in the sample. If a random sample of n observations, y1, y2, . . . , yn, is selected from a population that has a normal
distribution with mean and variance 2, then, by definition, the 2 sampling
distribution is
2 ⫽
S 2
2
(4.1)
where is the number of degrees of freedom in the sample and the other
terms are as defined previously.
4.2
DISTRIBUTIONS USED IN SAMPLING THEORY
53
A plot of the distribution is shown in Figure 4.1. The number of redundancies (degrees of freedom) in sample set statistics such as those for the
mean or variance are ⫽ n ⫺ 1; in later chapters on least squares it will be
shown that the number of redundancies is based on the number of independent
observations and unknown parameters. In the case of the mean, one observation is necessary for its determination, thus leaving n ⫺ 1 values as redundant observations. Table D.2 is a tabulation of 2 distribution curves that have
from 1 to 120 degrees of freedom. To find the area under the upper tail of
the curve (right side, shown hatched in Figure 4.1), we start at some specific
2 value and, going to infinity (⬁), intersect the row corresponding to the
appropriate degrees of freedom, , with the column corresponding to the
desired area under the curve. For example, to find the specific 2 value relating
to 1% (␣ ⫽ 0.010) of the area under a curve having 10 degrees of freedom,
we intersect the row headed by 10 with the column headed by 0.010 and find
a 2 value of 23.21. This means that 1% of the area under this curve is
between the values of 23.21 and ⬁.
Due to the asymmetric nature of the distribution shown in Figure 4.1, the
percentage points1 (␣) of the lower tail (left side of the curve) must be computed from those tabulated for the upper tail. A specific area under the left
side of the curve starting at zero and going to a specific 2 value is found by
subtracting the tabulated ␣ (right-side area) from 1. This can be done since
the table lists ␣ (areas) starting at the 2 value and going to ⬁, and the total
area under the curve is 1. For example, if there are 10 degrees of freedom
and the 2 value relating to 1% of the area under the left side of the curve is
needed, the row corresponding to equal to 10 is intersected with the column
headed by ␣ ⫽ 0.990 (1 ⫺ 0.010), and a value of 2.56 is obtained. This
means that 1% of the area under the curve occurs from 0 to 2.56.
The 2 distribution is used in sampling statistics to determine the range in
which the variance of the population can be expected to occur based on (1)
some specified percentage probability, (2) the variance of a sample set, and
(3) the number of degrees of freedom in the sample. In an example given in
Figure 4.1 2 distribution.
1
Percentage points are decimal equivalents of percent probability; that is, a percent probability of
95% is equivalent to 0.95 percentage points.
54
CONFIDENCE INTERVALS
Section 4.6, this distribution is used to construct probability statements about
the variance of the population being in a range centered about the variance
S2 of a sample having degrees of freedom. In Section 5.4 a statistical test
is presented using the 2 distribution to check if the variance of a sample is
a valid estimate for the population variance.
4.2.2 t (Student) Distribution
The t distribution is used to compare a population mean with the mean of a
sample set based on the number of redundancies () in the sample set. It is
similar to the normal distribution (discussed in Chapter 3) except that the
normal distribution applies to an entire population, whereas the t distribution
applies to a sampling of the population. The t distribution is preferred over
the normal distribution when the sample contains fewer than 30 values. Thus,
it is an important distribution in analyzing surveying data.
If z is a standard normal random variable as defined in Section 3.4, 2 is
a chi-square random variable with degrees of freedom, and z and 2 are
both independent variables, then by definition
t⫽
z
兹2 /
(4.2)
The t values for selected upper-tail percentage points (hatched area in Figure 4.2) versus the t distributions with various degrees of freedom are listed
in Table D.3. For specific degrees of freedom () and percentage points (␣),
the table lists specific t values that correspond to the areas ␣ under the curve
between the tabulated t values and ⬁. Similar to the normal distribution, the
t distribution is symmetric. Generally in statistics, only percentage points in
the range 0.0005 to 0.4 are necessary. These t values are tabulated in Table
D.3. To find the t value relating to ␣ ⫽ 0.01 for a curve developed with 10
degrees of freedom ( ⫽ 10), intersect the row corresponding to ⫽ 10 with
the row corresponding to ␣ ⫽ 0.01. At this intersection a t value of 2.764 is
obtained. This means that 1% (␣ ⫽ 0.01) of the area exists under the t distribution curve having 10 degrees of freedom in the interval between 2.764
and ⬁. Due to the symmetry of this curve, it can also be stated that 1% (␣
Figure 4.2 t distribution.
4.2
DISTRIBUTIONS USED IN SAMPLING THEORY
55
⫽ 0.01) of the area under the curve developed for 10 degrees of freedom also
lies between and 2.764.
As described in Section 4.3, this distribution is used to construct confidence
intervals for the population mean () based on the mean (y) and variance (S2)
of a sample set degrees of freedom. An example in that section illustrates
the procedure. Furthermore, in Section 5.3 it is shown that this distribution
can be used to develop statistical tests about the population mean.
4.2.3 F Distribution
The F distribution is used when comparing the variances computed from two
sample sets. If 12 and 22 are two chi-square random variables with 1 and 2
degrees of freedom, respectively, and both variables are independent, then by
definition
F
21 / 1
22 / 2
(4.3)
Various percentage points (areas under the upper tail of the curve shown
hatched in Figure 4.3) of the F distribution are tabulated in Table D.4. Notice
Figure 4.3 F distribution.
56
CONFIDENCE INTERVALS
that this distribution has 1 numerator degrees of freedom and 2 denominator
degrees of freedom, which correspond to the two sample sets. Thus, unlike
the 2 and t distributions, each desired ␣ percentage point must be represented
in a separate table. In Appendix D, tables for the more commonly used values
of ␣ (0.20, 0.10, 0.05, 0.025, 0.01, 0.005, and 0.001) are listed.
To illustrate the use of the tables, suppose that the F value for 1% of the
area under the upper tail of the curve is needed. Also assume that 5 is the
numerator degrees of freedom relating to S1 and 10 is the denominator degrees
of freedom relating to S2. In this example, ␣ equals 0.01 and thus the F table
in Table D.4 that is written for ␣ ⫽ 0.01 must be used. In that table, intersect
the row headed by 2 equal to 10 with the column headed by 1 equal to 5,
and find the F value of 5.64. This means that 1% of the area under the curve
constructed using these degrees of freedom lies in the region from 5.64 to ⬁.
To determine the area in the lower tail of this distribution, use the following
functional relationship:
F␣,1,2 ⫽
1
F1⫺␣,2,1
(4.4)
The critical F value for the data in the preceding paragraph [1 equal to 5
and 2 equal to 10 with ␣ equal to 0.99 (0.01 in the lower tail)] is determined
by going to the intersection of the row headed by 5 with the column headed
by 10 in the section ␣ ⫽ 0.01. The intersection is at F equal to 2.19. According to Equation (4.4), the critical F0.99,5,10 is 1/F0.01,10,5 ⫽ 1/2.19 ⫽ 0.457.
Thus, 1% of the area is under the F-distribution curve from to 0.457.
The F distribution is used to answer the question of whether two sample
sets come from the same population. For example, suppose that two samples
have variances of S 21 and S 22. If these two sample variances represent the same
population variance, the ratio of their population variances (21 / 22) should
equal 1 (i.e., 21 22). As discussed in Section 4.7, this distribution enables
confidence intervals to be established for the ratio of the population variances.
Also, as discussed in Section 5.5, the distribution can be used to test whether
the ratio of the two variances is statistically equal to 1.
4.3
CONFIDENCE INTERVAL FOR THE MEAN: t STATISTIC
In Chapter 3 the standard normal distribution was used to predict the range
in which the mean of a population can exist. This was based on the mean
and standard deviation for a sample set. However, as noted previously, the
normal distribution is based on an entire population, and as was demonstrated,
variations from the normal distribution are expected from sample sets having
a small number of values. From this expectation, the t distribution was developed. As demonstrated later in this section by an example, the t distribution
4.3
CONFIDENCE INTERVAL FOR THE MEAN: t STATISTIC
57
(in Table D.3) for samples having an infinite number of values uses the same
t values as those listed in Table 3.2 for the normal distribution. It is generally
accepted that when the number of observations is greater than about 30, the
values in Table 3.2 are valid for constructing intervals about the population
mean. However, when the sample set has fewer than 30 values, a t value from
the t distribution should be used to construct the confidence interval for the
population mean.
To derive an expression for a confidence interval of the population mean,
a sample mean (y) is computed from a sample set of a normally distributed
population having a mean of and a variance in the mean of 2 /n. Let z ⫽
(y ⫺ )/( / 兹n) be a normal random variable. Substituting it and Equation
(4.1) into Equation (4.2) yields
t⫽
z
兹 /
2
⫽
(y ⫺ )/( / 兹n)
兹(S / )/
2
2
⫽
(y ⫺ )/( / 兹n) y ⫺
⫽
S/
S/ 兹n
(4.5)
To compute a confidence interval for the population mean () given a
sample set mean and variance, it is necessary to determine the area of a 1 ⫺
␣ region. For example, in a 95% confidence interval (nonhatched area in
Figure 4.4), center the percentage point of 0.95 on the t distribution. This
leaves 0.025 in each of the upper- and lower-tail areas (hatched areas in Figure
4.4). The t value that locates an ␣ /2 area in both the upper and lower tails
of the distribution is given in Table D.3 as t␣ / 2,. For sample sets having a
mean of y and a variance of S2, the correct probability statement to locate
this area is
P(z ⬍ t) ⫽ 1 ⫺ ␣
Substituting Equation (4.5) into Equation (a) yields
P
y⫺
⬍t ⫽1⫺␣
S/ n
which after rearranging yields
Figure 4.4 t␣ / 2 plot.
(a)
58
CONFIDENCE INTERVALS
P y ⫺ t␣ / 2
S
S
⬍ ⬍ y ⫹ t␣ / 2
⫽1⫺␣
n
n
(4.6)
Thus, given y, t␣ / 2,, n, and S, it is seen from Equation (4.6) that a 1 ⫺ ␣
probable error interval for the population mean is computed as
y ⫺ t␣ / 2
S
S
⬍ ⬍ y ⫹ t␣ / 2
n
n
(4.7)
where t␣ / 2 is the t value from the t distribution based on degrees of freedom
and ␣ /2 percentage points.
The following example illustrates the use of Equation (4.7) and Table D.3
for determining the 95% confidence interval for the population mean based
on a sample set having a small number of values (n) with a mean of y and a
variance of S.
Example 4.1 In carrying out a control survey, 16 directional readings were
measured for a single line. The mean (seconds’ portion only) of the readings
was 25.4ⴖ, with a standard deviation of 1.3ⴖ. Determine the 95% confidence
interval for the population mean. Compare this with the interval determined
by using a t value determined from the standard normal distribution tables
(Table 3.2).
SOLUTION In this example the confidence level 1 ⫺ ␣ is 0.95, and thus ␣
is 0.05. Since the interval is to be centered about the population mean , a
value of ␣ /2 in Table D.3 is used. This yields equal areas in both the lower
and upper tails of the distribution, as shown in Figure 4.4. Thus, for this
example, ␣ /2 is 0.025. The appropriate t value for this percentage point with
equal to 15 (16 ⫺ 1) degrees of freedom is found in Table D.3 as follows:
Step 1: In the leftmost column of Table D.3, find the row with the correct
number of degrees of freedom () for the sample. In this case it is 16 ⫺
1, or 15.
Step 2: Find the column headed by 0.025 for ␣ /2.
Step 3: Locate the value at the intersection of this row and column, which is
2.131.
Step 4: Then by Equation (4.7), the appropriate 95% confidence interval is
24.7 ⫽ 25.4 ⫺ 2.131
⬍ ⬍ y ⫹ t0.025
1.3
16
⫽ y ⫺ t0.025
S
n
S
1.3
⫽ 25.4 ⫹ 2.131
⫽ 26.1
n
16
4.4
TESTING THE VALIDITY OF THE CONFIDENCE INTERVAL
59
This computation can be written more compactly as
y t0.025
S
or
n
25.4 2.131
1.3
⫽ 25.4 0.7
16
After making the calculation above, it can be stated that for this sample,
with 95% confidence, the population mean () lies in the range (24.7, 26.1).
If this were a large sample, the t value from Table 3.2 could be used for 95%.
That t value for 95% is 1.960, and the standard error of the mean then would
be 1.3/ 16 ⫽ 0.325. Thus, the population’s mean would be in the range
25.4 1.960 ⫻ 0.325ⴖ, or (24.8, 26.0). Notice that due to the small sample
size, the t distribution gives a larger range for the population mean than does
the standard normal distribution. Notice also that in the t distribution of Table
D.3, for a sample of infinite size (i.e., ), the t value tabulated for ␣
equal to 0.025 is 1.960, which matches Table 3.2.
The t distribution is often used to isolate outliers or blunders in observations. To do this, a percent confidence interval is developed about the mean
for a single observation as
y t␣ / 2S ⱕ yi ⱕ y t␣ / 2S
(4.8)
Using the data from Example 4.1 and Equation (4.8), the 95% range for the
16 directional readings is
25.4ⴖ 2.131(1.3ⴖ) 22.63ⴖ ⱕ yi ⱕ 28.17ⴖ 25.4ⴖ 2.131(1.3ⴖ)
Thus, 95% of the data should be in the range (22.6ⴖ, 28.2ⴖ). Any data values
outside this range can be considered as outliers and rejected with a 95% level
of confidence. It is important to note that if the normal distribution value of
1.960 was used to compute this interval, the range would be smaller, (22.85ⴖ,
27.95ⴖ). Using the normal distribution could result in discarding more observations than is justified when using sample estimates of the mean and variance. It is important to note that this will become more significant as the
number of observations in the sample becomes smaller. For example, if only
four directional readings are obtained, the t-distribution multiplier would become 3.183. The resulting 95% confidence interval for a single observation
would be 1.6 times larger than that derived using a normal distribution t value.
4.4
TESTING THE VALIDITY OF THE CONFIDENCE INTERVAL
A test that demonstrates the validity of the theory of the confidence interval
is illustrated as follows. Using a computer and normal random number gen-
60
CONFIDENCE INTERVALS
erating software, 1000 sample data sets of 16 values each were collected
randomly from a population with mean ⫽ 25.4 and standard error ⫽
1.3. Using a 95% confidence interval (␣ ⫽ 0.05) and Equation (4.7), the
interval for the population mean derived for each sample set was computed
and compared with the actual population mean. If the theory is valid, the
interval constructed would be expected to contain the population’s mean 95%
of the time based on the confidence level of 0.05. Appendix E shows the 95%
intervals computed for the 1000 samples. The intervals not containing the
population mean of 25.4 are marked with an asterisk. From the data tabulated
it is seen that 50 of 1000 sample sets failed to contain the population mean.
This corresponds to exactly 5% of the samples. In other words, the proportion
of samples that do enclose the mean is exactly 95%. This demonstrates that
the bounds calculated by Equation (4.7) do, in fact, enclose the population
mean at the confidence level selected.
4.5
SELECTING A SAMPLE SIZE
A common problem encountered in surveying practice is to determine the
number of repeated observations necessary to meet a specific precision. In
practice, the size of S cannot be controlled absolutely. Rather, as seen in
Equation (4.7), the confidence interval can be controlled only by varying the
number of repeated observations. In general, the larger the sample size, the
smaller the confidence interval. From Equation (4.7), the range in which the
population mean () resides at a selected level of confidence (␣) is
y t␣ / 2
S
n
(b)
Now let I represent one-half of the interval in which the population mean
lies. Then from Equation (b), I is
S
n
(4.9)
(4.10)
I ⫽ t␣ / 2
Rearranging Equation (4.9) yields
n⫽
t␣ / 2S
I
2
In Equation (4.10), n is the number of repeated measurements, I the desired
confidence interval, t␣ / 2 the t value based on the number of degrees of freedom
(), and S the sample set standard deviation. In the practical application of
Equation (4.10), t␣ / 2 and S are unknown since the data set has yet to be
4.6
CONFIDENCE INTERVAL FOR A POPULATION VARIANCE
61
collected. Also, the number of measurements, and thus the number of redundancies, is unknown, since they are the computational objectives in this problem. Therefore, Equation (4.10) must be modified to use the standard normal
random variable, z, and its value for t, which is not dependent on or n; that
is,
2
t
n ⫽ ␣/2
I
(4.11)
where n is the number of repetitions, t␣ / 2 the t value determined from the
standard normal distribution table (Table D.1), an estimated value for the
standard error of the measurement, and I the desired confidence interval.
Example 4.2 From the preanalysis of a horizontal control network, it is
known that all angles must be measured to within 2ⴖ at the 95% confidence
level. How many repetitions will be needed if the standard deviation for a
single angle measurement has been determined to be 2.6ⴖ?
SOLUTION In this problem, a final 95% confidence interval of 2ⴖ is
desired. From previous experience or analysis,2 the standard error for a single
angle observation is estimated as 2.6ⴖ. From Table 3.2, the multiplier (or t
value) for a 95% confidence level is found to be 1.960. Substituting this into
Equation (4.11) yields
n⫽
1.960 ⫻ 2.6
2
2
⫽ 6.49
Thus, eight repetitions are selected, since this is the closest even number
above 6.49. [Note that to eliminate instrumental systematic errors it is necessary to select an even number of repetitions, because an equal number of
face-left (direct) and face-right (reverse) readings must be taken.]
4.6
CONFIDENCE INTERVAL FOR A POPULATION VARIANCE
From Equation (4.1), 2 ⫽ S 2 / 2, and thus confidence intervals for the variance of the population, 2, are based on the 2 statistic. Percentage points
(areas) for the upper and lower tails of the 2 distribution are tabulated in
Table D.2. This table lists values (denoted by 2␣) that determine the upper
boundary for areas from 2␣ to ⫹⬁ of the distribution, such that
2
See Chapter 6 for a methodology to estimate the variance in an angle observation.
62
CONFIDENCE INTERVALS
P(2 ⬎ ␣2 ) ⫽ ␣
for a given number of redundancies, . Unlike the normal distribution and the
t distribution, the 2 distribution is not symmetric about zero. To locate an
area in the lower tail of the distribution, the appropriate value of 21⫺␣ must
be found, where P(2 ⬎ 21⫺␣) ⫽ 1 ⫺ ␣. These facts are used to construct a
probability statement for 2 as
P(21⫺␣ / 2 ⬍ 2 ⬍ 2␣ / 2) ⫽ 1 ⫺ ␣
(4.12)
where 21⫺␣ / 2 and 2␣ / 2 are tabulated in Table D.2 by the number of redundant
observations. Substituting Equation (4.1) into Equation (4.12) yields
S 2
⬍ 2␣ / 2
2
P 21⫺␣ / 2 ⬍
⫽P
21⫺␣ / 2
1
2
⬍ 2 ⬍ ␣ /22
2
S
S
(4.13)
Recalling a property of mathematical inequalities—that in taking the reciprocal of a function, the inequality is reversed—it follows that
P
S 2
S 2
2
⬍
2␣ / 2
21␣ / 2
1␣
(4.14)
Thus, the 1 ␣ confidence interval for the population variance (2) is
S 2
S 2
2
2␣ / 2
21␣ / 2
(4.15)
Example 4.3 An observer’s pointing and reading error with a 1ⴖ theodolite
is estimated by collecting 20 readings while pointing at a well-defined distant
target. The sample standard deviation is determined to be 1.8ⴖ. What is the
95% confidence interval for 2?
SOLUTION For this example the desired area enclosed by the confidence
interval 1 ␣ is 0.95. Thus, ␣ is 0.05 and ␣ /2 is 0.025. The values of
20.025 and 20.975 with equal to 19 degrees of freedom are needed. They are
found in the 2 table (Table D.2) as follows:
Step 1: Find the row with 19 degrees of freedom and intersect it with the
column headed by 0.975. The value at the intersection is 8.91.
Step 2: Follow this procedure for 19 degrees of freedom and 0.025. The value
is 32.85. Using Equation (4.15), the 95% confidence interval for 2 is
4.7
CONFIDENCE INTERVAL FOR THE RATIO OF TWO POPULATION VARIANCES
63
(20 ⫺ 1)1.82
(20 ⫺ 1)1.82
⬍ 2 ⬍
32.85
8.91
1.87 ⬍ 2 ⬍ 6.91
Thus, 95% of the time, the population’s variance should lie between 1.87
and 6.91.
4.7 CONFIDENCE INTERVAL FOR THE RATIO OF TWO
POPULATION VARIANCES
Another common statistical procedure is used to compare the ratio of two
population variances. The sampling distribution of the ratio 21 / 22 is well
known when samples are collected randomly from a normal population. The
confidence interval for 21 / 22 is based on the F distribution using Equation
(4.3) as
F⫽
21 / 1
22 / 2
Substituting Equation (4.1) and reducing yields
F⫽
(1S 12 / 12)/ 1 S 12 / 21 S 21 22
⫽
⫽
(2S 22 / 22)/ 2 S 22 / 22 S 22 21
(4.16)
To establish a confidence interval for the ratio, the lower and upper values
corresponding to the tails of the distribution must be found. A probability
statement to find the confidence interval for the ratio is constructed as follows:
P(F1⫺␣ / 2,1,2 ⬍ F ⬍ F␣ / 2,1,2) ⫽ 1 ⫺ ␣
Substituting in Equation (4.16) and rearranging yields
P(Fl ⬍ F ⬍ Fu) ⫽ P Fl ⬍
⫽ P Fl
⫽P
S 21 22
⫻
⬍ Fu
S 22 12
S 22 22 S 22
⬍
⬍
F
S 12 12 S 12 u
1 S 12 12 S 12 1
⬍
⬍
⫽1⫺␣
Fu S 22 22 S 22 Fl
Substituting Equation (4.4) into (4.17) yields
(4.17)
64
CONFIDENCE INTERVALS
P
1
F␣ / 2,1,2
⫽P
S 21 21 S 21
1
⬍
⬍
S 22 22 S 22 F1⫺␣ / 2,12
1
F␣ / 2,1,2
S 21 21 S 21
⬍
⬍
F
⫽1⫺␣
S 22 22 S 22 ␣ / 2,2,1
(4.18)
Thus, from Equation (4.18), the 1 ⫺ ␣ confidence interval for the 21 / 22
ratio is
1
F␣ / 2,12
S 21 21 S 21
⬍
⬍
F
S 22 22 S 22 ␣ / 2,2,1
(4.19)
Notice that the degrees of freedom for the upper and lower limits in Equation
(4.19) are opposite each other, and thus 2 is the numerator degrees of freedom and 1 is the denominator degrees of freedom in the upper limit.
An important situation where Equation (4.19) can be applied occurs in the
analysis and adjustment of horizontal control surveys. During least squares
adjustments of these types of surveys, control stations fix the data in space
both positionally and rotationally. When observations tie into more than a
minimal number of control stations, the control coordinates must be mutually
consistent. If they are not, any attempt to adjust the observations to the control
will warp the data to fit the discrepancies in the control. A method for isolating control stations that are not consistent is first to do a least squares
adjustment using only enough control to fix the data in space both positionally
and rotationally. This is known as a minimally constrained adjustment. In
horizontal surveys, this means that one station must have fixed coordinates
and one line must be fixed in direction. This adjustment is then followed with
an adjustment using all available control. If the control information is consistent, the reference variance (S 21) from the minimally constrained adjustment
should be statistically equivalent to the reference variance (S 22) obtained when
using all the control information. That is, the ratio of S 21 / S 22 should be 1.
Example 4.4 Assume that a minimally constrained trilateration network adjustment with 24 degrees of freedom has a reference variance of 0.49 and
that the fully constrained network adjustment with 30 degrees of freedom has
a reference variance of 2.25. What is the 95% 1 ⫺ ␣ confidence interval for
the ratio of the variances and does this interval contain the numerical value
1? Stated in another way, is there reason to be concerned about the control
having values that are not consistent?
SOLUTION In this example the objective is to determine whether the two
reference variances are statistically equal. To solve the problem, let the variance in the numerator be 2.25 and that in the denominator be 0.49. Thus,
the numerator has 30 degrees of freedom (1 ⫽ 30) and corresponds to an
PROBLEMS
65
adjustment using all the control. The denominator has 24 degrees of freedom
(2 ⫽ 24) and corresponds to the minimally constrained adjustment.3 With ␣
equal to 0.05 and using Equation (4.19), the 95% confidence interval for this
ratio is
2.08
2.25
1
2
2.25
12
(2.14) 9.83
0.49 2.21
2
0.49
Note from the calculations above that 95% of the time, the ratio of the
population variances is in the range (2.08, 9.83). Since this interval does not
contain 1, it can be said that 21 / 22 1 and 21 22 at a 95% level of
confidence. Recalling from Equation (2.4) that the size of the variance depends on the size of the errors, it can be stated that the fully constrained
adjustment revealed discrepancies between the observations and the control.
This could be caused by inconsistencies in the coordinates of the control
stations or by the presence of uncorrected systematic errors in the observations. An example of an uncorrected systematic error is the failure to reduce
distance observations to a mapping grid before the adjustment. (See Appendix
F.4 for a discussion of the reduction of distance observations to the mapping
grid.)
PROBLEMS
4.1
Use the 2-distribution table (Table D.2) to determine the values of
2␣ / 2 that would be used to construct confidence intervals for a population variance for the following combinations:
(a) ␣ 0.10, 25
(b) ␣ 0.05, 15
(c) ␣ 0.05, 10
(d) ␣ 0.01, 30
4.2
Use the t-distribution table (Table D.3) to determine the values of t␣ / 2
that would be used to construct confidence intervals for a population
mean for each of the following combinations:
(a) ␣ 0.10, 25
(b) ␣ 0.05, 15
(c) ␣ 0.01, 10
(d) ␣ 0.01, 40
3
For confidence intervals, it is not important which variance is selected as the numerator. In this
case, the larger variance was selected arbitrarily as the numerator, to match statistical testing
methods discussed in Chapter 5.
66
CONFIDENCE INTERVALS
4.3
Use the F-distribution table (Table D.4) to determine the values of
F␣,1,2 that would be used to construct confidence intervals for a population mean for each of the following combinations:
(a) ␣ ⫽ 0.20, 1 ⫽ 24, 2 ⫽ 2
(b) ␣ ⫽ 0.01, 1 ⫽ 15, 2 ⫽ 8
(c) ␣ ⫽ 0.05, 1 ⫽ 60, 2 ⫽ 20
(d) ␣ ⫽ 0.80, 1 ⫽ 2, 2 ⫽ 24
Use STATS to do Problems 4.4 through 4.6.
4.4
Problem 4.1
4.5
Problem 4.2
4.6
Problem 4.3
4.7
A least squares adjustment is computed twice on a data set. When the
data are minimally constrained with 10 degrees of freedom, a variance
of 1.07 is obtained. In the second run, the fully constrained network
has 12 degrees of freedom with a standard deviation of 1.53. The a
priori estimates for the reference variances in both adjustments are 1;
that is, 21 ⫽ 22 ⫽ 1.
(a) What is the 95% confidence interval for the ratio of the two variances? Is there reason to be concerned about the consistency of
the control? Justify your response statistically.
(b) What is the 95% confidence interval for the reference variance in
the minimally constrained adjustment? The population variance is
1. Does this interval contain 1?
(c) What is the 95% confidence interval for the reference variance in
the fully constrained adjustment? The population variance is 1.
Does this interval contain 1?
4.8
The calibrated length of a baseline is 402.167 m. An average distance
of 402.151 m with a standard deviation of 0.0055 m is computed
after the line is observed five times with an EDM.
(a) What is the 95% confidence interval for the measurement?
(b) At a 95% level of confidence, can you state that the EDM is working properly? Justify your response statistically.
(c) At a 90% level of confidence can you state that the EDM is working properly? Justify your response statistically.
4.9
An observer’s pointing and reading standard deviation is determined
to be 1.8ⴖ after pointing and reading the circles of a particular instrument six times (n ⫽ 6). What is the 99% confidence interval for
the population variance?
PROBLEMS
67
4.10 Using sample statistics and the data in Example 3.1, construct a 90%
confidence interval:
(a) for a single observation, and identify any observations that may be
identified as possible outliers.
(b) for the population variance.
4.11 Using sample statistics and the data in Example 3.2, construct a 90%
confidence interval:
(a) for a single observation, and identify any observations that may be
identified as possible outliers.
(b) for the population variance.
4.12 Using sample statistics and the data from Problem 3.10 construct a
95% confidence interval:
(a) for a single observation, and identify any observations that may be
identified as possible outliers.
(b) for the population variance.
4.13 For the data in Problem 3.11, construct a 95% confidence interval for
the ratio of the two variances for sets 1 and 2. Are the variances equal
statistically at this level of confidence? Justify your response statistically.
4.14 Using sample statistics and the data from Problem 3.12, construct a
95% confidence interval:
(a) for a single observation, and identify any observations that may be
identified as possible outliers in the data.
(b) for the mean.
CHAPTER 5
STATISTICAL TESTING
5.1
HYPOTHESIS TESTING
In Example 4.4 we were not concerned about the actual bounds of the interval
constructed, but rather, whether the interval contained the expected ratio of
the variances. This is often the case in statistics. That is, the actual values of
the interval are not as important as is answering the question: Is the sample
statistic consistent with what is expected from the population? The procedures
used to test the validity of a statistic are known as hypothesis testing. The
basic elements of hypothesis testing are
1. The null hypothesis, H0, is a statement that compares a population statistic with a sample statistic. This implies that the sample statistic is
what is ‘‘expected’’ from the population. In Example 4.4, this would be
that the ratio of the variances is statistically 1.
2. The alternative hypothesis, Ha, is what is accepted when a decision is
made to reject the null hypothesis, and thus represents an alternative
population of data from which the sample statistic was derived. In Example 4.4 the alternative hypothesis would be that the ratio of the variances is not equal to 1 and thus the variance did not come from the
same population of data.
3. The test statistic is computed from the sample data and is the value
used to determine whether the null hypothesis should be rejected. When
the null hypothesis is rejected, it can be said that the sample statistic
computed is not consistent with what is expected from the population.
In Example 4.4 a rejection of the null hypothesis would occur when the
ratio of the variances is not statistically equivalent to 1.
68
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
5.1
HYPOTHESIS TESTING
69
4. The rejection region is the value for the test statistic where the null
hypothesis is rejected. In reference to confidence intervals, this number
takes the place of the confidence interval bounds. That is, when the test
statistic computed is greater than the value defining the rejection region,
it is equivalent to the sample statistic of the null hypothesis being outside the bounds of the confidence interval. That is, when the rejection
criterion is true, the null hypothesis is rejected.
Whenever a decision is made concerning the null hypothesis, there is a
possibility of making a wrong decision since we can never be 100% certain
about a statistic or a test. Returning to Example 4.4, a confidence interval of
95% was constructed. With this interval, there is a 5% chance that the decision
was wrong. That is, it is possible that the larger-than-expected ratio of the
variances is consistent with the population of observations. This reasoning
suggests that further analysis of statistical testing is needed.
Two basic errors can occur when a decision is made about a statistic. A
valid statistic could be rejected, or an invalid statistic could be accepted. These
two errors can be stated in terms of statistical testing elements as Type I and
Type II errors. If the null hypothesis is rejected when in fact it is true, a Type
I error is committed. If the null hypothesis is not rejected when in fact it is
false, a Type II error occurs. Since these errors are not from the same population, the probability of committing each error is not directly related. A
decision must be made as to the type of error that is more serious for the
situation, and the decision should be based on the consequences of committing each error. For instance, if a contract calls for positional accuracies on
95% of the stations to be within 0.3 ft, the surveyor is more inclined to
commit a Type I error to ensure that the contract specifications are met.
However, the same surveyor, needing only 1-ft accuracy on control to support
a small-scale mapping project, may be more inclined to commit a Type II
error. In either case, it is important to compute the probabilities of committing
both Type I and Type II errors to assess the reliability of the inferences derived
from a hypothesis test. For emphasis, the two basic hypothesis-testing errors
are repeated.
•
•
Type I error: rejecting the null hypothesis when it is, in fact, true (symbolized by ␣)
Type II error: not rejecting the null hypothesis when it is, in fact, false
(symbolized by )
Table 5.1 shows the relationship between the decision, the probabilities of
␣ and , and the acceptance or rejection of the null hypothesis, H0. In Figure
5.1 the left distribution represents the data from which the null hypothesis is
derived. That is, this distribution represents a true null hypothesis. Similarly,
the distribution on the right represents the distribution of data for the true
alternative. These two distributions could be attributed to measurements that
70
STATISTICAL TESTING
TABLE 5.1 Relationships in Statistical Testing
Decision
Situation
H0 true
H0 false (Ha true)
Accept H0
Reject H0
Correct decision: P ⫽ 1 ⫺ ␣
(confidence level)
Type II error: P ⫽ 
Type I error: P ⫽ ␣
(significance level)
Correct decision: P ⫽ 1 ⫺ 
(power of test)
contain only random errors (left distribution) versus measurements containing
blunders (right distribution). In the figure it is seen that valid measurements
in the ␣ region of the left distribution are being rejected at a significance
level of ␣. Thus, ␣ represents the probability of committing a Type I error.
This is known as the significance level of the test. Furthermore, data from the
right distribution are being accepted at a  level of significance. The power
of the test is 1 ⫺  and corresponds to a true alternative hypothesis. Methods
of computing  or 1 ⫺  are not clear, or are often difficult, since nothing is
generally known about the distribution of the alternative. Consequently, in
statistical testing, the objective is to prove the alternative hypothesis true by
showing that the data do not support the statistic coming from the null hypothesis distribution. In doing this, only a Type I error can be made, for which
a known probability of making an incorrect decision is ␣.
Example 5.1 Assume that for a population of 10,000 people, a flu virus test
has a 95% confidence level and thus a significance level, ␣, of 0.05. Suppose
that 9200 people test negative for the flu virus and 800 test positive. Of the
800 people who tested positive, 5%, or 40 people, will test incorrectly (false
positive). That is, they will test positive for the flu but do not have it. This is
an example of committing a Type I error at an ␣ level of significance. Similarly assume that 460 people test negative for the flu when, in fact, they do
have it (a false-negative case). This is an example of a Type II error at a
Figure 5.1 Graphical interpretation of Type I and Type II errors.
5.2
SYSTEMATIC DEVELOPMENT OF A TEST
71
probability of , that is equal to 0.046 (460/10,000). The power of the test
is 1 ⫺  or 0.954.
From the foregoing it is seen that it is possible to set the probability of
committing a Type I error for a given H0. However, for a fixed level of ␣ and
sample size n, the probability of a Type II error, , is generally unknown. If
the null hypothesis, H0, and ␣ are fixed, the power of the test can only be
increased by increasing the sample size, n. Since the power of the test may
be low or unknown, statisticians always say that the test failed to reject the
null hypothesis, rather than making any statements about its acceptance. This
is an important statistical concept. That is, it should never be stated that the
null hypothesis is accepted since the power of the test is unknown. It should
only be said that ‘‘there is no statistical evidence to reject the null hypothesis.’’
Because of this small but important distinction, it is important to construct a
test that rejects the null hypothesis whenever possible.
A similar situation exists with surveying measurements. If a distance measurement contains a large systematic error, it is possible to detect this with a
fully constrained adjustment and thus reject the null hypothesis. However, if
a distance contains a very small systematic error, the ability to detect the
systematic error may be low. Thus, although some confidence can be placed
in the rejection of the null hypothesis, it can never be said that the null
hypothesis should be accepted since the probability of undetected small systematic errors or blunders cannot be determined. What we strive to do is
minimize the size of these errors so that they have little effect on the computed
results.
5.2
SYSTEMATIC DEVELOPMENT OF A TEST
When developing a statistical test, the statistician must determine the test
variables and the type of test to perform. This book will look at statistical
tests for the mean, variance, and ratio of two sample variances. The t test is
used when comparing a sample mean versus a population mean. This test
compares the mean of a set of observations against a known calibrated value.
The 2 test is used when comparing a sample variance against a population
variance. As discussed in Section 16.7, this test is used in a least squares
adjustment when comparing the reference variance from an adjustment
against its population value. Finally, when comparing variances from two
different sample sets, the F test is used. As discussed in Section 21.6, this
test is used in least squares adjustments when comparing the reference variances from a minimally constrained and fully constrained adjustment. Table
5.2 lists the test variables of these three statistical tests.
72
STATISTICAL TESTING
TABLE 5.2 Test Variables and Statistical Tests
Variable 1,
Test Statistic
Population mean,
Population variance, 2
Ratio of sample
variances equals 1
Variable 2,
Sample Statistic
Null Hypothesis
Test Statistic
Sample mean, y
Sample variance, S2
S21 / S22
H0: ⫽ y
H0: 2 ⫽ S2
H0: S21 / S22 ⫽ 1
t
2
F
A test can take two forms based on the distributions. A one-tailed test uses
the critical value from either the left or right side of a distribution, whereas
the two-tailed test is much like a confidence interval, with the critical value
divided equally on both sides of the distribution.
In the one-tailed test, the concern is whether the sample statistic is either
greater or less than the statistic being tested. In the two-tailed test, the concern
is whether the sample statistic is different from the statistic being tested. For
example, when checking the angle-reading capabilities of a total station
against the manufacturer’s specifications, a surveyor would not be concerned
if the instrument were working at a level better than the manufacturer’s stated
accuracy. However, the surveyor would probably send the instrument in for
repairs if it was performing at a level below the manufacturer’s stated accuracy. In this case, it would be appropriate to perform a one-tailed test. On the
other hand, when checking the mean distance observed using an EDM against
a known calibration baseline length, the surveyor wants to know if the mean
length is statistically different from the calibrated length. In this case it is
appropriate to perform a two-tailed test. In the following sections it is important (1) identify the appropriate test statistic and (2) the type of test to
perform.
In all forms of statistical testing, a test statistic is developed from the data.
The test statistic is then compared against a critical value from the distribution. If the rejection region statement is true, the null hypothesis is rejected
at the level of significance selected. As stated earlier, this is the goal of a
well-developed test since only Type I error occurs at the selected level of
significance. If the rejection region statement is false, the test fails to reject
null hypothesis. Because of the possibility of Type II error and due to the
lack of knowledge about the alternative distribution, no statement about the
validity of the null hypothesis can be made; at best it can be stated that there
is no reason to reject the null hypothesis.
5.3
TEST OF HYPOTHESIS FOR THE POPULATION MEAN
At times it may be desirable to test a sample mean against a known value.
The t distribution is used to build this test. The null hypothesis for this test
can take two forms: one- or two-tailed tests. In the one-tailed test, the concern
5.3
73
TEST OF HYPOTHESIS FOR THE POPULATION MEAN
is whether the sample mean is either statistically greater or less than the
population mean. In the two-tailed test, the concern is whether the sample
mean is statistically different from the population mean. These two tests are
shown below.
One-Tailed Test
Two-Tailed Test
Null hypothesis:
Ha: ⫽ y
H0: ⫽ y
Alternative hypothesis:
Ha: ⬎ y( ⬍ y)
Ha: ⫽ y
The test statistic is
t⫽
y⫺
S/ n
(5.1)
The region where the null hypothesis is rejected is
t ⬎ t␣ (or t ⬍ t␣)
t ⬎ t␣ / 2
It should be stated that for large samples (n ⬎ 30), the t value can be
replaced by the standard normal variate, z.
Example 5.2 A baseline of calibrated length 400.008 m is observed repeatedly with an EDM instrument. After 20 observations, the average of the observed distances is 400.012 m with a standard deviation of 0.002 m. Is the
distance observed significantly different from the distance calibrated at a 0.05
level of significance?
SOLUTION Assuming that proper field and office procedures were followed, the fundamental question is whether the EDM is working within its
specifications and thus providing distance observations in a population of
calibrated values. To answer this question, a two-tailed test is used to determine whether the distance is the same or is different from the distance calibrated at a 0.05 level of significance. That is, the mean of the distances
observed will be rejected if it is statistically either too short or too long to
be considered the same as the calibration value. The rationale behind using
a two-tailed test is similar to that used when constructing a confidence interval, as in Example 4.1. That is, 2.5% of the area from the lower and upper
tails of the t distribution is to be excluded from the interval constructed, or
in this case, the test.
The null hypothesis is
H0: ⫽ 400.012
and the alternative hypothesis is
74
STATISTICAL TESTING
Ha: ⫽ 400.012
By Equation (5.1), the test statistic is
t⫽
y ⫺ 400.012 ⫺ 400.008
⫽
⫽ 8.944
S/ n
0.002/ 20
and the rejection region is
t ⫽ 8.944 ⬎ t␣ / 2
Since a two-tailed test is being done, the ␣ /2 (0.025) column in the tdistribution table is intersected with the ⫽ n ⫺ 1, or 19 degrees of freedom,
row. From the t distribution (Table D.3), t0.025, 19 is found to be 2.093, and
thus the rejection region is satisfied. In other words, the value computed for
t is greater than the tabulated value, and thus the null hypothesis can be
rejected at a 95% level of confidence. That is,
t ⫽ 8.944 ⬎ t␣ / 2 ⫽ 2.093
Based on the foregoing, there is reason to believe that the average length
observed is significantly different from its calibrated value at a 5% significance level. This implies that at least 5% of the time, the decision will be
wrong. As stated earlier, a 95% confidence interval for the population mean
could also have been constructed to derive the same results. Using Equation
(4.7), that interval would yield
冉 冊
冉 冊
400.011 ⫽ 400.012 ⫺ 2.093
0.002
兹20
ⱕ ⱕ 400.012 ⫹ 2.093
0.002
⫽ 400.013
兹20
Note that the 95% confidence interval fails to contain the baseline value of
400.008, and similarly, there is reason to be concerned about the calibration
status of the instrument. That is, it may not be working properly and should
be repaired.
5.4
TEST OF HYPOTHESIS FOR THE POPULATION VARIANCE
In Example 5.2, the procedure for checking whether an observed length compares favorably with a calibrated value was discussed. The surveyor may also
5.4
TEST OF HYPOTHESIS FOR THE POPULATION VARIANCE
75
want to check if the instrument is measuring at its published precision. The
2 distribution is used when comparing the variance of a sample set against
that of a population. This test involves checking the variance computed from
a sample set of observations against the published value (the expected variance of the population).
As shown in Table 5.2, the 2 distribution checks the sample variance
against a population variance. By using Equation (4.1), the following statistical test is written.
One-Tailed Test
Two-Tailed Test
Null hypothesis:
H0: S2 ⫽ 2
H0: S2 ⫽ 2
Alternative hypothesis:
Ha: S2 ⬎ 2 (or Ha: S2 ⬍ 2)
Ha: S2 ⫽ 2
The test statistic is
2 ⫽
S 2
2
(5.2)
from which the null hypothesis is rejected when the following statement is
satisfied:
2 ⬎ 2␣
(or 2 ⬍ 21⫺␣)
2 ⬍ 21⫺␣ / 2
(or 2 ⬎ 2␣ / 2)
The rejection region is determined from Equation (4.13). Graphically, the
null hypothesis is rejected in the one-tailed test when the 2 value computed
is greater than the value tabulated. This rejection region is the shaded region
shown in Figure 5.2(a). In the two-tailed test, the null hypothesis is rejected
when the value computed is either less than 21⫺␣ / 2 or greater than 2␣ / 2. This
is similar to the computed variance being outside the constructed confidence
interval for the population variance. Again in the two-tailed test, the proba-
Figure 5.2 Graphical interpretation of (a) one- and (b) two-tailed tests.
76
STATISTICAL TESTING
bility selected is evenly divided between the upper and lower tails of the
distribution such that the acceptance region is centered on the distribution.
These rejection regions are shown graphically in Figure 5.2(b).
Example 5.3 The owner of a surveying firm wants all surveying technicians
to be able to read a particular instrument to within 1.5ⴖ. To test this value,
the owner asks the senior field crew chief to perform a reading test with the
instrument. The crew chief reads the circle 30 times and obtains r ⫽ 0.9ⴖ.
Does this support the 1.5ⴖ limit at a 5% level of significance?
SOLUTION In this case the owner wishes to test the hypothesis that the
computed sample variance is the same as the population variance rather than
being greater than the population variance. That is, all standard deviations
that are equal to or less than 1.5ⴖ will be accepted. Thus, a one-tailed test is
constructed as follows (note that ⫽ 30 ⫺ 1, or 29): The null hypothesis is
H0: S2 ⫽ 2
and the alternative hypothesis is
Ha: S2 ⬎ 2
The test statistic is
2 ⫽
(30 ⫺ 1)0.92
⫽ 10.44
1.52
The null hypothesis is rejected when the computed test statistic exceeds the
tabulated value, or when the following statement is true:
2 ⫽ 10.44 ⬎ 2␣, ⫽ 20.05,29 ⫽ 42.56
where 42.56 is from Table D.2 for 20.05,29. Since the computed 2 value (10.44)
is less than the tabulated value (42.56), the null hypothesis cannot be rejected.
However, simply failing to reject the null hypothesis does not mean that
the value of 1.5ⴖ is valid. This example demonstrates a common problem
in statistical testing when results are interpreted incorrectly. A valid sample
set from the population of all surveying employees cannot be obtained
by selecting only one employee. Furthermore, the test is flawed since every
instrument reads differently and thus new employees may initially have
problems reading an instrument, due to their lack of experience with the
instrument. To account properly for this lack of experience, the employer
could test a random sample of prospective employees during the interview
process, and again after several months of employment. The owner could then
check for a correlation between the company’s satisfaction with the employee,
5.5
TEST OF HYPOTHESIS FOR THE RATIO OF TWO POPULATION VARIANCES
77
and the employee’s initial ability to read the instrument. However, it is unlikely that any correlation would be found. This is an example of misusing
statistics.
Example 5.3 illustrates an important point to be made when using statistics.
The interpretation of statistical testing requires judgment by the person performing the test. It should always be remembered that with a test, the objective is to reject and not accept the null hypothesis. Furthermore, a statistical
test should be used only where appropriate.
5.5 TEST OF HYPOTHESIS FOR THE RATIO OF TWO
POPULATION VARIANCES
When adjusting data, surveyors have generally considered control to be absolute and without error. However, like any other quantities derived from
observations, it is a known fact that control may contain error. As discussed
in Example 4.4, one method of detecting both errors in control and possible
systematic errors in horizontal network measurements is to do both a minimally constrained and a fully constrained least squares adjustment with the
data. After doing both adjustments, the post-adjustment reference variances
can be compared. If the control is without error and no systematic errors are
present in the data, the ratio of the two reference variances should be close
to 1. Using Equation (4.18), a hypothesis test can be constructed to compare
the ratio of variances for two sample sets as follows:
One-Tailed Test
Two-Tailed Test
Null hypothesis:
H0:
S 21
⫽1
S 22
(i.e., S 21 ⫽ S 22)
H0:
S 21
⫽1
S 22
(i.e., S 21 ⫽ S 22)
(i.e., S 21 ⬎ S 22)
Ha:
S 21
⫽1
S 22
(i.e., S 21 ⫽ S 22)
Alternative hypothesis:
Ha:
S 21
⬎1
S 22
or
Ha:
S 21
⬍1
S 22
(i.e., S 12 ⬍ S 22)
The test statistic that will be used to determine rejection of the null hypothesis
is
78
STATISTICAL TESTING
F⫽
S 21
S 22
or F ⫽
S 22
S 21
F⫽
larger sample variance
smaller sample variance
The null hypothesis should be rejected when the following statement is
satisfied:
F ⬎ F␣ / 2
F ⬎ F␣
F␣ and F␣ / 2 are values that locate the ␣ and ␣ /2 areas, respectively, in the
upper tail of the F distribution with 1 numerator degrees of freedom and 2
denominator degrees of freedom. Notice that in the two-tailed test, the degrees
of freedom of the numerator are taken from the numerically larger sample
variance, and the degrees of freedom of the denominator are from the smaller
variance.
Example 5.4 Using the same data as presented in Example 4.4, would the
null hypothesis be rejected?
SOLUTION In this example, a two-tailed test is appropriate since the only
concern is whether the two reference variances are equal statistically. In this
problem, the interval is centered on the F distribution with an ␣ /2 area in the
lower and upper tails. In the analysis, 30 degrees of freedom in the numerator
correspond to the larger sample variance, and 24 degrees of freedom in the
denominator to the smaller variance, so that the following test is constructed:
The null hypothesis is
H0:
S 21
⫽1
S 22
Ha:
S 21
⫽1
S 22
and the alternative hypothesis is
The test statistic for checking rejection of the null hypothesis is
F⫽
2.25
⫽ 4.59
0.49
Rejection of the null hypothesis occurs when the following statement is true:
F ⫽ 4.59 ⬎ F␣ / 2,1,2 ⫽ F0.025,30,24 ⫽ 2.21
5.5
TEST OF HYPOTHESIS FOR THE RATIO OF TWO POPULATION VARIANCES
79
Here it is seen that the F value computed (4.59) is greater than its value
from Table D.4 (2.21). Thus, the null hypothesis can be rejected. In other
words, the fully constrained adjustment does not have the same variance as
its minimally constrained counterpart at the 0.05 level of significance. Notice
that the same result was obtained here as was obtained in Example 4.4 with
the 95% confidence interval. Again, the network should be inspected for the
presence of systematic errors, followed by an analysis of possible errors in
the control stations. This post-adjustment analysis is revisited in greater detail
in Chapter 20.
Example 5.5 Ron and Kathi continually debate who measures angles more
precisely with a particular total station. After listening to enough of this debate, their supervisor describes a test where each is to measure a particular
direction by pointing and reading the instrument 51 times. They must then
compute the variance for their data. At the end of the 51 readings, Kathi
determines her variance to be 0.81 and Ron finds his to be 1.21. Is Kathi a
better instrument operator at a 0.01 level of significance?
SOLUTION In this situation, even though Kathi’s variance implies that her
observations are more precise than Ron’s, a determination must be made to
see if the reference variances are statistically equal versus Kathi’s being better
than the Ron’s. This test requires a one-tailed F test with a significance level
of ␣ ⫽ 0.01. The null hypothesis is
H0:
S 2R
⫽1
S 2K
(S 2R ⫽ S 2K)
and the alternative hypothesis is
Ha:
S 2R
⬎1
S 2K
(S 2R ⬎ S 2K)
The test statistic is
F⫽
1.21
⫽ 1.49
0.81
The null hypothesis is rejected when the computed value for F (1.49) is
greater than the tabulated value for F0.01,50,50 (1.95). Here it is seen that the
value computed for F is less than its tabulated value, and thus the test statistic
does not satisfy the rejection region. That is,
80
STATISTICAL TESTING
F ⫽ 1.49 ⬎ F␣,50,50 ⫽ 1.95 is false
Therefore, there is no statistical reason to believe that Kathi is better than
Ron at a 0.01 level of significance.
Example 5.6 A baseline is observed repeatedly over a period of time using
an EDM instrument. Each day, 10 observations are taken and averaged. The
variances for the observations are listed below. At a significance level of 0.05,
are the results of day 2 significantly different from those of day 5?
Day
Variance, S 2 (mm2)
1
2
3
4
5
50.0
61.0
51.0
53.0
54.0
SOLUTION This problem involves checking whether the variances of days
2 and 5 are statistically equal or are different. This is the same as constructing
a confidence interval involving the ratio of the variances. Because the concern
is about equality or inequality, this will require a two-tailed test. Since 10
observations were collected each day, both variances are based on 9 degrees
of freedom (1 and 2). Assume that the variance for day 2 is S 22 and the
variance for day 5 is S 25. The test is constructed as follows: The null hypothesis is
H0:
S 22
⫽1
S 25
Ha:
S 22
⫽1
S 25
and the alternative hypothesis is
The test statistic is
F⫽
61
⫽ 1.13
54
The null hypothesis is rejected when the computed F value (1.13) is greater
than the value in Table D.4 (4.03). In this case the rejection region is F ⫽
1.13 ⬎ F0.025,9,9 ⫽ 4.03 and is not satisfied. Consequently, the test fails to
reject the null hypothesis, and there is no statistical reason to believe that the
data of day 2 are statistically different from those of day 5.
PROBLEMS
81
PROBLEMS
5.1
In your own words, state why the null hypothesis can never be
accepted.
5.2
Explain why medical tests on patients are performed several times in
a laboratory before the results of the test are returned to the doctor.
5.3
In your own words, discuss when it is appropriate to use:
(a) a one-tailed test.
(b) a two-tailed test.
5.4
Match the following comparison with the appropriate test.
(a) A calibration baseline length against a value measured using an
EDM.
(b) Two sample variances.
(c) The reference variance of a fully constrained adjustment against a
minimally constrained adjustment.
(d) The reference variance of a least squares adjustment against its a
priori value of 1.
5.5
Compare the variances of days 2 and 5 in Example 5.6 at a level of
significance of 0.20 (␣ ⫽ 0.20). Would testing the variances of days 1
and 2 result in a different finding?
5.6
Compare the variances of days 1 and 4 in Example 5.6 at a level
significance of 0.05 (␣ ⫽ 0.05). Would testing the variances of days 1
and 3 result in a different finding?
5.7
Using the data given in Example 5.5, determine if Kathi is statistically
better with the equipment than Ron at a significance level of:
(a) 0.05.
(b) 0.10.
5.8
The population value for the reference variance from a properly
weighted least squares adjustment is 1. After running a minimally constrained adjustment having 15 degrees of freedom, the reference variance is computed 1.52. Is this variance statistically equal to 1 at:
(a) a 0.01 level of significance?
(b) a 0.05 level of significance?
(c) a 0.10 level of significance?
5.9
When all the control is added to the adjustment in Problem 5.8, the
reference variance for the fully constrained adjustment with 15 degrees
of freedom is found to be 1.89. Are the reference variances from the
minimally constrained and fully constrained adjustments statistically
equal at:
82
STATISTICAL TESTING
(a) a 0.01 level of significance?
(b) a 0.05 level of significance?
(c) a 0.10 level of significance?
5.10 The calibrated length of a baseline is 402.267 m. A mean observation
for the distance is 402.251 m with a standard deviation of 0.0052 m
after six readings with an EDM.
(a) Is the distance observed statistically different from the length calibrated at a 5% level of significance?
(b) Is the distance observed statistically different from the length calibrated at a 10% level of significance?
5.11 A mean length of 1023.573 m with a standard deviation of 0.0056
m is obtained for a distance after five observations. Using the technical
specifications, it is found that the standard deviation for this observation should be 0.0043 m.
(a) Perform a statistical test to check the repeatability of the instrument
at a level of significance of 0.05.
(b) Perform a statistical test to check the repeatability of the instrument at a level of significance of 0.01.
5.12 A least squares adjustment is computed twice on a data set. When the
data are minimally constrained with 24 degrees of freedom, a reference
variance of 0.89 is obtained. In the second run, the fully constrained
network, which also has 24 degrees of freedom, has a reference variance of 1.15. The a priori estimate for the reference variance in both
adjustments is 1; that is, 21 ⫽ 22 ⫽ 1.
(a) Are the two variances statistically equal at a 0.05 level of
significance?
(b) Is the minimally constrained adjustment reference variance statistically equal to 1 at a 0.05 level of significance?
(c) Is the fully constrained adjustment reference variance statistically
equal to 1 at a 0.05 level of significance?
(d) Is there a statistical reason to be concerned about the presence of
errors in either the control or the observations?
5.13 A total station with a manufacturer’s specified angular accuracy of 5ⴖ
was used to collect the data in Problem 5.12. Do the data warrant this
accuracy at a 0.05 level of significance? Develop a statistical test to
validate your response.
5.14 A surveying company decides to base a portion of their employees’
salary raises on improvement in their use of equipment. To determine
the improvement, the employees measure their ability to point on a
target and read the circles of a theodolite every six months. One em-
PROBLEMS
83
ployee is tested six weeks after starting employment and obtains a
standard deviation of 1.5ⴖ with 25 measurements. Six months later
the employee obtains a standard deviation of 1.2ⴖ with 30 measurements. Did the employee improve statistically over six months at a 5%
level of significance? Is this test an acceptable method of determining
improvements in quality? What suggestion, if any, would you give to
modify the test?
5.15 An EDM is placed on a calibration baseline and the distance between
two monuments is determined to be 1200.012 m 0.047 m after 10
observations. The length between the monuments is calibrated as
1200.005 m. Is the instrument measuring the length properly at:
(a) a 0.01 level of significance?
(b) a 0.05 level of significance?
(c) a 0.10 level of significance?
CHAPTER 6
PROPAGATION OF RANDOM
ERRORS IN INDIRECTLY
MEASURED QUANTITIES
6.1
BASIC ERROR PROPAGATION EQUATION
As discussed in Section 1.2, unknown values are often determined indirectly
by making direct measurements of other quantities which are functionally
related to the desired unknowns. Examples in surveying include computing
station coordinates from distance and angle observations, obtaining station
elevations from rod readings in differential leveling, and determining the azimuth of a line from astronomical observations. As noted in Section 1.2, since
all quantities that are measured directly contain errors, any values computed
from them will also contain errors. This intrusion, or propagation, of errors
that occurs in quantities computed from direct measurements is called error
propagation. This topic is one of the most important discussed in this book.
In this chapter it is assumed that all systematic errors have been eliminated,
so that only random errors remain in the direct observations. To derive the
basic error propagation equation, consider the simple function, z ⫽ a1x1 ⫹
a2x2, where x1 and x2 are two independently observed quantities with standard
errors 1 and 2, and a1 and a2 are constants. By analyzing how errors propagate in this function, a general expression can be developed for the propagation of random errors through any function.
Since x1 and x2 are two independently observed quantities, they each have
different probability density functions. Let the errors in n determinations of
x1 be εi1, εii1, . . . , εn1 and the errors in n determinations of x2 be εi2, εii2, . . . ,
εn2; then zT , the true value of z for each independent measurement, is
84
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
6.1
85
BASIC ERROR PROPAGATION EQUATION
冦
a1(x1i ⫺ ε1i ) ⫹ a2(x2i ⫺ ε2i ) ⫽ a1x1i ⫹ a2xi2 ⫺ (a1ε1i ⫹ a2εi2)
zT ⫽
a1(x1ii ⫺ ε1ii) ⫹ a2(x2ii ⫺ ε2ii) ⫽ a1x1ii ⫹ a2x2ii ⫺ (a1ε1ii ⫹ a2ε2ii)
a1(x1iii ⫺ ε1iii) ⫹ a2(x2iii ⫺ ε2iii) ⫽ a1x1iii ⫹ a2x2iii ⫺ (a1ε1iii ⫹ a2ε2iii)
(6.1)
⯗
The values for z computed from the observations are
zi ⫽ a1xi1 ⫹ a2xi2
zii ⫽ a1xii1 ⫹ a2xii2
(6.2)
ziii ⫽ a1x1iii ⫹ a2xiii
2
⯗
Substituting Equations (6.2) into Equations (6.1) and regrouping Equations
(6.1) to isolate the errors for each computed value yields
zi ⫺ zT ⫽ a1ε1i ⫹ a2ε2i
zii ⫺ zT ⫽ a1ε1ii ⫹ a2ε2ii
iii
iii
1 1
(6.3)
iii
2 2
z ⫺ zT ⫽ a ε ⫹ a ε
⯗
From Equation (2.4) for the variance in a population, n2 ⫽ 兺ni⫽1 ε2, and
thus for the case under consideration, the sum of the squared errors for the
value computed is
冘 ε ⫽ (a ε ⫹ a ε ) ⫹ (a ε
n
2
i
i
1 1
i 2
2 2
ii
1 1
⫹ a2ε2ii)2 ⫹ (a1ε1iii ⫹ a2ε2iii)2 ⫹ 䡠 䡠 䡠 ⫽ nz2
i⫽1
(6.4)
Expanding the terms in Equation (6.4) yields
n2z ⫽ (a1ε1i )2 ⫹ 2a1a2ε1i ε2i ⫹ (a2ε2i )2 ⫹ (a1ε1ii)2 ⫹ 2a1a2ε1iiε2ii ⫹ (a2ε2ii)2 ⫹ 䡠 䡠 䡠
(6.5)
Factoring terms in Equation (6.5) gives
2
2
2
2
2
2
n2z ⫽ a21(εi1 ⫹ εii1 ⫹ ε1iii ⫹ 䡠 䡠 䡠) ⫹ a22(εi2 ⫹ ε2ii ⫹ ε2iii ⫹ 䡠 䡠 䡠)
⫹ 2a1a2(ε1i ε2i ⫹ ε1iiε2ii ⫹ ε1iiiε2iii ⫹ 䡠 䡠 䡠)
(6.6)
Inserting summation symbols for the error terms in Equation (6.6) results in
86
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
2
z
2
i
⫽a
冉
兺ni⫽1 ε12
n
冊
⫹ 2a1a2
冉
兺ni⫽1 ε1ε2
n
冊 冉
⫹a
兺ni⫽1 ε22
2
2
n
冊
(6.7)
Recognizing that the terms in parentheses in Equation (6.7) are by definition:
2x1, x1x2, and x22, respectively, Equation (6.7) can be rewritten as
2z ⫽ a122x1 ⫹ 2a1a2x1x2 ⫹ a222x2
(6.8)
In Equation (6.8), the middle term, x1x2, is known as the covariance. This
term shows the interdependence between the two unknown variables, x1 and
x2. As the covariance term decreases, the interdependence of the variables
also decreases. When these terms are zero, the variables are said to be mathematical independent. Its importance is discussed in more detail in later
chapters.
Equations (6.7) and (6.8) can be written in matrix form as
冋
⌺zz [a1
a2]
2x1
x1x2
x1x2
x22
册冋 册
a1
a2
(6.9)
where ⌺zz is the variance–covariance matrix for the function z. It follows
logically from this derivation that, in general, if z is a function of n independently measured quantities, x1, x2, . . . , xn, then ⌺zz is
⌺zz ⫽ [a1
冤
2x1
an] x2x1
⯗
xnx1
a2 䡠 䡠 䡠
x1x2
2x2
xnx2
䡠䡠䡠
䡠䡠䡠
冥冤 冥
x1xn
x2xn
a1
a2
(6.10)
⯗
2xn
an
Further, for a set of m functions with n independently measured quantities,
x1, x2, . . . , xn, Equation (6.10) expands to
⌺zz ⫽
冤
a11
a21
a12
a22
⯗
⯗
am1 am2
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠 a1n
䡠 a2n
䡠 ⯗
䡠 amn
冥冤
x21 x1x2
x1x2 x22
⯗
⯗
xnx1 x2xn
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠 x1xn
䡠 x2xn
⯗
䡠
䡠 x2n
冥冤
冥
a11 a21 䡠 䡠 䡠 am1
a12 a22 䡠 䡠 䡠 am2
⯗
⯗
䡠䡠䡠
⯗
a1n a2n 䡠 䡠 䡠 amn
(6.11)
Similarly, if the functions are nonlinear, a first-order Taylor series expansion can be used to linearize them.1 Thus, a11, a12, . . . are replaced by the
partial derivatives of Z1, Z2, . . . with respect to the unknown parameters, x1,
1
Readers who are unfamiliar with solving nonlinear equations should refer to Appendix C.
6.1
87
BASIC ERROR PROPAGATION EQUATION
x2, . . . . Therefore, after linearizing a set of nonlinear equations, the matrix
for the function of Z can be written in linear form as
⌺zz ⫽
⭸Z1
⭸x1
⭸Z2
⭸x1
⯗
⭸Zm
⭸x1
⭸Z1
⭸x2
⭸Z2
⭸x2
⯗
⭸Zm
⭸x2
⭸Z1
⭸xn
⭸Z2
䡠䡠䡠
⭸xn
䡠䡠䡠 ⯗
⭸Zm
䡠䡠䡠
⭸xn
䡠䡠䡠
x21 x1x2
x1x2 2x2
⯗
⯗
xnx1 x2xn
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠 x1xn
䡠 x2xn
䡠 ⯗
䡠 x2n
⭸Z1
⭸x1
⭸Z1
⭸x2
⯗
⭸Z1
⭸xn
⭸Z2
⭸x1
⭸Z2
⭸x2
⯗
⭸Z2
⭸xn
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
⭸Zm
⭸x1
⭸Zm
⭸x2
⯗
⭸Zm
⭸xn
(6.12)
Equations (6.11) and (6.12) are known as the general law of propagation
of variances (GLOPOV) for linear and nonlinear equations, respectively.
Equations (6.11) and (6.12) can be written symbolically in matrix notation as
⌺zz ⫽ A⌺AT
(6.13)
where ⌺zz is the covariance matrix for the function Z. For a nonlinear set of
equations that is linearized using Taylor’s theorem, the coefficient matrix (A)
is called a Jacobian matrix, a matrix of partial derivatives with respect to
each unknown, as shown in Equation (6.12).
If the measurements x1, x2, . . . , xn are unrelated (i.e., are statistically
independent), the covariance terms x1x2, x1x3, . . . equal to zero, and thus the
right-hand sides of Equations (6.10) and (6.11) can be rewritten, respectively,
as
⌺zz ⫽
a11
a21
a12
a22
⯗
⯗
am1
am2
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
a1n
a2n
⯗
amn
x21
0
⯗
x22
⯗
0
0
0
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
0
0
⯗
x2n
a11 a21
a12 a22
⯗
⯗
a1n a2n
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
am1
am2
⯗
amn
(6.14)
⌺zz ⫽
⭸Z1
⭸x1
⭸Z2
⭸x1
⯗
⭸Zm
⭸x1
⭸Z1
⭸x2
⭸Z2
⭸x2
⯗
⭸Zm
⭸x2
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
⭸Z1
⭸xn
⭸Z2
⭸xn
⯗
⭸Zm
⭸xn
2x1
0
⯗
2x2
⯗
0
0
0
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
0
0
⯗
2xn
⭸Z1
⭸x1
⭸Z1
⭸x2
⯗
⭸Z1
⭸xn
⭸Z2
⭸x1
⭸Z2
⭸x2
⯗
⭸Z2
⭸xn
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
⭸Zm
⭸x1
⭸Zm
⭸x2
⯗
⭸Zm
⭸xn
(6.15)
88
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
If there is only one function Z, involving n unrelated quantities, x1, x2,
. . . , xn, Equation (6.15) can be rewritten in algebraic form as
Z ⫽
冪冉
冊 冉 冊
⭸Z
⭸x1 x1
2
⭸Z
⭸x2 x2
⫹
2
⫹䡠䡠䡠⫹
冉 冊
⭸Z
⭸xn xn
2
(6.16)
Equations (6.14), (6.15), and (6.16) express the special law of propagation
of variances (SLOPOV). These equations govern the manner in which errors
from statistically independent measurements (i.e., xixj ⫽ 0) propagate in a
function. In these equations, individual terms (⭸Z/ ⭸xi)xi represent the individual contributions to the total error that occur as the result of observational
errors in each independent variable. When the size of a function’s estimated
error is too large, inspection of these individual terms will indicate the largest
contributors to the error. The most efficient way to reduce the overall error
in the function is to closely examine ways to reduce the largest error terms
in Equation (6.16).
6.1.1
Generic Example
Let A ⫽ B ⫹ C, and assume that B and C are independently observed quantities. Note that ⭸A/ ⭸B ⫽ 1 and ⭸Z/ ⭸C ⫽ 1. Substituting these into Equation
(6.16) yields
A ⫽ 兹(1B)2 ⫹ (1C)2
(6.17)
Using Equation (6.15) yields
⌺AA ⫽ [1
冋
1]
册冋 册
2B
0
0
2C
1
⫽ [B2 ⫹ C2 ]
1
Equation (6.17) yields the same results as Equation (6.16) after the square
root of the single element is determined. In the equations above, standard
error () and standard deviation (S) can be used interchangeably.
6.2
6.2.1
FREQUENTLY ENCOUNTERED SPECIFIC FUNCTIONS
Standard Deviation of a Sum
Let A ⫽ B1 ⫹ B2 ⫹ 䡠 䡠 䡠 ⫹ Bn, where the B’s are n independently observed
quantities having standard deviations of SB1, SB2, . . . , SBn. Then by Equation
(6.16),
6.3
SA ⫽
6.2.2
NUMERICAL EXAMPLES
S 2B1 ⫹ S 2B2 ⫹ 䡠 䡠 䡠 ⫹ S B2 n
89
(6.18)
Standard Deviation in a Series
Assume that the error for each observed value in Equation (6.18) is equal;
that is, SB1, SB2, . . . , SBn ⫽ SB then Equation (6.18) simplifies to
(6.19)
SA ⫽ SB n
6.2.3
Standard Deviation of the Mean
Let y be the mean obtained from n independently observed quantities y1, y2,
. . . , yn, each of which has the same standard deviation S. As given in
Equation (2.1), the mean is expressed as
y⫽
y1 ⫹ y2 ⫹ 䡠 䡠 䡠 ⫹ yn
n
An equation for Sy, the standard deviation of y, is obtained by substituting
the expression above into Equation (6.16). Since the partial derivatives of y
with respect to the observed quantities, y1, y2, . . . , yn, is ⭸y / ⭸y1 ⫽ ⭸y / ⭸y2 ⫽
䡠 䡠 䡠 ⫽ ⭸y / ⭸yn ⫽ 1/n, the resulting error in y is
Sy ⫽
冪冉
1
S
n y1
冊 冉 冊
2
⫹
1
S
n y2
2
⫹䡠䡠䡠⫹
冉 冊
1
S
n yn
2
⫽
冪
nS 2
S
⫽
2
n
兹n
(6.20)
Note that Equation (6.20) is the same as Equation (2.8).
6.3
NUMERICAL EXAMPLES
Example 6.1 The dimensions of the rectangular tank shown in Figure 6.1
are measured as
L ⫽ 40.00 ft
SL ⫽ Ⳳ0.05 ft
W ⫽ 20.00 ft
SW ⫽ Ⳳ0.03 ft
H ⫽ 15.00 ft
SH ⫽ Ⳳ0.02 ft
Find the tank’s volume and the standard deviation in the volume using the
measurements above.
90
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
Figure 6.1 Rectangular tank.
SOLUTION The volume of the tank is found using the formula
V ⫽ LWH ⫽ 40.00(20.00)(15.00) ⫽ 12,000 ft3
Given that ⭸V/ ⭸L ⫽ WH, ⭸V/ ⭸W ⫽ LH, and ⭸V/ ⭸H ⫽ LW, the standard
deviation in the volume is determined by using Equation (6.16), which yields
SV ⫽
冪冉
冊 冉 冊 冉 冊
⭸V
S
⭸L L
2
⭸V
S
⭸W W
⫹
2
⫹
⭸V
S
⭸H H
2
⫽ 兹(WH)2(0.05)2 ⫹ (LH)2(0.03)2 ⫹ (LW)2(0.02)2
(a)
⫽ 兹(300 ⫻ 0.05) ⫹ (600 ⫻ 0.03) ⫹ (800 ⫻ 0.02)
2
2
2
⫽ 兹225 ⫹ 324 ⫹ 256 ⫽ 兹805 ⫽ Ⳳ28 ft3
In Equation (a), the second term is the largest contributor to the total error,
and thus to reduce the overall error in the computed volume, it would be
prudent first to try to make SW smaller. This would yield the greatest effect
in the error of the function.
Example 6.2 As shown in Figure 6.2, the vertical angle ␣ to point B is
observed at point A as 3⬚00⬘, with S␣ being Ⳳ1⬘. The slope distance D from
A to B is observed as 1000.00 ft, with SD being Ⳳ0.05 ft. Compute the
horizontal distance and its standard deviation.
Figure 6.2 Horizontal distance from slope observations.
6.3
91
NUMERICAL EXAMPLES
SOLUTION The horizontal distance is determined using the equation
H ⫽ D cos ␣ ⫽ 1000.00 cos(3⬚00⬘) ⫽ 998.63 ft
Given that ⭸H/ ⭸D ⫽ cos ␣ and ⭸H/ ⭸␣ ⫽ ⫺D sin ␣, the error in the function
is determined by using Equation (6.16) as
SH ⫽
冪冉 冊 冉 冊
⭸H
S
⭸D D
2
⫹
⭸H
S
⭸␣ ␣
2
(6.21)
In Equation (6.21), S␣ must be converted to its equivalent radian value to
achieve agreement in the units. Thus,
SH ⫽
冪
⫽
冪
冉
冉
冊
60ⴖ
(cos ␣ ⫻ 0.05) ⫹ ⫺sin ␣ ⫻ D ⫻
206,264.8ⴖ /rad
2
(0.9986 ⫻ 0.05)2 ⫹
冊
⫺0.0523 ⫻ 1000 ⫻ 60ⴖ
206,264.8ⴖ
2
2
⫽ 兹0.049932 ⫹ 0.01522 ⫽ Ⳳ0.052 ft
Notice in this example that the major contributing error source (largest
number under the radical) is 0.049932. This is the error associated with the
distance measurement, and thus if the resulting error of Ⳳ0.052 ft is too large,
the logical way to improve the results (reduce the overall error) is to adopt a
more precise method of measuring the distance.
Example 6.3 The elevation of point C on the chimney shown in Figure 6.3
is desired. Field angles and distances are observed. Station A has an elevation
of 1298.65 Ⳳ 0.006 ft, and station B has an elevation of 1301.53 Ⳳ 0.004
Figure 6.3 Elevation of a chimney determined using intersecting angles.
92
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
ft. The instrument height, hiA, at station A is 5.25 Ⳳ 0.005 ft, and the instrument height, hiB, at station B is 5.18 Ⳳ 0.005 ft. The other observations and
their errors are
AB ⫽ 136.45 Ⳳ 0.018
A ⫽ 44⬚12⬘34ⴖ Ⳳ 8.6ⴖ
B ⫽ 39⬚26⬘56ⴖ Ⳳ 11.3ⴖ
v1 ⫽ 8⬚12⬘47ⴖ Ⳳ 4.1ⴖ
v2 ⫽ 5⬚50⬘10ⴖ Ⳳ 5.1ⴖ
What are the elevation of the chimney and the error in this computed value?
SOLUTION Normally, this problem is worked in several steps. The steps
include computing distances AI and BI and then solving for the average elevation of C using observations obtained from stations A and B. However,
caution must be exercised when doing error analysis in a stepwise fashion
since the computed values could be correlated and the stepwise method might
lead to an incorrect analysis of the errors. To avoid this, either GLOPOV can
be used or a single function can be derived that includes all quantities observed when the elevation is calculated. The second method is demonstrated
as follows.
From the sine law, the solution of AI and BI can be derived as
AI ⫽
AB sin B
AB sin B
⫽
sin[180⬚ ⫺ (A ⫹ B)] sin(A ⫹ B)
(6.22)
BI ⫽
AB sin A
sin(A ⫹ B)
(6.23)
Using Equations (6.22) and (6.23), the elevations for C from stations A and
B are
ElevCA ⫽ AI tan v1 ⫹ ElevA ⫹ hiA
(6.24)
ElevCB ⫽ BI tan v2 ⫹ ElevB ⫹ hiB
(6.25)
Thus, the chimney’s elevation is computed as the average of Equations (6.24)
and (6.25), or
ElevC ⫽ –21 (ElevCA ⫹ ElevCB)
(6.26)
Substituting Equations (6.22) through (6.25) into (6.26), a single expression
for the chimney elevation can be written as
6.3
ElevC
⫽
93
NUMERICAL EXAMPLES
1
AB sin B tan v1
AB sin A tan v2
ElevA ⫹ hiA ⫹
⫹ ElevB ⫹ hiB ⫹
2
sin(A ⫹ B)
sin(A ⫹ B)
(6.27)
From Equation (6.27), the elevation of C is 1316.49 ft. To perform the
error analysis, Equation (6.16) is used. In this complex problem, it is often
easier to break the problem into smaller parts. This can be done by numerically solving each partial derivative necessary for Equation (6.16) before
squaring and summing the results. From Equation (6.26),
⭸ElevC
⭸ElevC
1
⫽
⫽
⭸ElevA
⭸ElevB
2
⭸ElevC
⭸ElevC
1
⫽
⫽
⭸hiA
⭸hiB
2
From Equation (6.27),
⭸ElevC
1
⫽
⭸AB
2
冉
⭸ElevC
AB
⫽
⭸A
2
冊
sin B tan v1 ⫹ sin A tan v2
⫽ 0.08199
sin(A ⫹ B)
冋
册
⫺cos(A ⫹ B)(sin B tan v1 ⫹ sin A tan v2)
cos A ⫹ tan v2
⫹
sin2(A ⫹ B)
sin(A ⫹ B)
⫽ 3.78596
⭸ElevC
AB
⫽
⭸B
2
冋
册
⫺cos(A ⫹ B)(sin B tan v1 ⫹ sin A tan v2)
cos B ⫹ tan v1
⫹
sin2(A ⫹ B)
sin(A ⫹ B)
⫽ 6.40739
⭸ElevC
AB sin B
⫽
⫽ 44.52499
⭸v1
2 sin(A ⫹ B)cos2v1
⭸ElevC
AB sin A
⫽
⫽ 48.36511
⭸v2
2 sin(A ⫹ B)cos2v2
Again for compatibility of the units in this problem, all angular errors are
converted to their radian equivalents by dividing each by 206,264.8ⴖ /rad.
Finally, using Equation (6.16), the error in the elevation computed is
94
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
S
2
ElevC
⫽
冉
冊 冉
冊 冉
冊 冉
⭸ElevC
S
⭸ElevA ElevA
⫹
⫹
冉
冉
2
⫹
2
⭸ElevC
S
⭸hiB hiB
⭸ElevC
SB
⭸B
⫹
2
SElevC
冉 冊 冉 冊 冉
⫽
冤
0.006
2
2
⫹
0.004
2
2
⫹2
⫹
冊 冉
冊 冉
冊 冉
⭸ElevC
S
⭸ElevB ElevB
⭸ElevC
S
⭸AB AB
⭸ElevC
Sv1
⭸v1
冊
1
0.005
2
2
⫹
2
⫹
2
⫹
冊
⭸ElevC
S
⭸hiA hiA
冊
冊
2
⭸ElevC
SA
⭸A
⭸ElevC
Sv2
⭸v2
2
2
2
冥
1/2
⫹ (0.08199 ⫻ 0.018)2
⫹ (3.78596 ⫻ 4.1693 ⫻ 10⫺5)2 ⫹ (6.40739 ⫻ 5.4783 ⫻ 10⫺5)2
⫹ (44.52499 ⫻ 1.9877 ⫻ 10⫺5)2 ⫹ (48.36511 ⫻ 2.4725 ⫻ 10⫺5)2
⫽ Ⳳ0.0055 ft ⬇ Ⳳ0.01 ft
Thus, the elevation of point C is 1316.49 Ⳳ 0.01 ft.
6.4
CONCLUSIONS
Errors associated with any indirect measurement problem can be analyzed as
described above. Besides being able to compute the estimated error in a function, the sizes of the individual errors contributing to the functional error can
also be analyzed. This identifies those observations whose errors are most
critical in reducing the functional error. An alternative use of the error propagation equation involves computing the error in a function of observed values
prior to fieldwork. The calculation can be based on the geometry of the problem and the observations that are included in the function. The estimated
errors in each value can be varied to correspond with those expected using
different combinations of available equipment and field procedures. The particular combination that produces the desired accuracy in the final computed
function can then be adopted in the field. This analysis falls under the heading
of survey planning and design. This topic is discussed further in Chapters 7
and 19.
The computations in this chapter can be time consuming and tedious, often
leading to computational errors in the results. It is often more efficient to
program these equations in a computational package. The programming of
the examples in this chapter is demonstrated in the electronic book on the
CD that accompanies this book.
PROBLEMS
95
PROBLEMS
6.1
In running a line of levels, 18 instrument setups are required, with a
backsight and foresight taken from each. For each rod reading, the
error estimated is Ⳳ0.005 ft. What is the error in the measured elevation difference between the origin and the terminus?
6.2
In Problem 2.6, compute the estimated error in the overall distance as
measured by both the 100- and 200-ft tapes. Which tape produced the
smallest error?
6.3
Determine the estimated error in the length of AE, which was measured
in sections as follows:
Section
AB
BC
CD
DE
Measured Length (ft)
Standard Deviation (ft)
416.24
1044.16
590.03
714.28
Ⳳ0.02
Ⳳ0.05
Ⳳ0.03
Ⳳ0.04
6.4
A slope distance is observed as 1506.843 Ⳳ 0.009 m. The zenith angle
is observed as 92⬚37⬘29ⴖ Ⳳ 8.8ⴖ. What is the horizontal distance and
its uncertainty?
6.5
A rectangular parcel has dimensions of 538.056 Ⳳ 0.005 m by 368.459
Ⳳ 0.004 m. What is the area of the parcel and the uncertainty in this
area?
6.6
The volume of a cone is given by V ⫽ –12 D2h. The cone’s measured
height is 8.5 in., with Sh ⫽ Ⳳ0.15 in. Its measured diameter is 5.98
in., with SD ⫽ Ⳳ0.05 in. What are the cone’s volume and standard
deviation?
6.7
An EDM instrument manufacturer publishes the instrument’s accuracy
as Ⳳ(3 mm ⫹ 3 ppm). [Note: 3 ppm means 3 parts per million. This
is a scaling error and is computed as (distance ⫻ 3/1,000,000).]
(a) What formula should be used to determine the error in a distance
observed with this instrument?
(b) What is the error in a 1864.98-ft distance measured with this
EDM?
6.8
As shown in Figure P6.8, a racetrack is measured in three simple components: a rectangle and two semicircles. Using an EDM with a manufacturer’s specified accuracy of Ⳳ(5 mm ⫹ 5 ppm), the rectangle’s
dimensions measured at the inside of the track are 5279.95 ft by 840.24
ft. Assuming only errors in the distance observations, what is:
96
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
(a)
(b)
(c)
(d)
(e)
the area enclosed by the track?
the length of the track?
the standard deviation in each track dimension?
the standard deviation in the perimeter of the track?
the standard deviation in the area enclosed by the track?
Figure P6.8
6.9
Using an EDM instrument, the rectangular dimensions of a large building 1435.67 Ⳳ 0.025 ft by 453.67 Ⳳ 0.01 ft are laid out. Assuming
only errors in distance observations, what is:
(a) the area enclosed by the building and its standard deviation?
(b) the perimeter of the building and its standard deviation?
6.10 A particular total station’s reading error is determined to be Ⳳ2.5ⴖ.
After pointing repeatedly on a distant target with the same instrument,
an observer determines an error due to both pointing and reading the
circles of Ⳳ3.6ⴖ. What is the observer’s pointing error?
6.11 For each tape correction formula noted below, express the error propagation formula in the form of Equation (6.16) using the variables
listed.
(a) H ⫽ L cos ␣, where L is the slope length and ␣ is the slope angle.
Determine the error with respect to L and ␣.
(b) CT ⫽ k(Tƒ ⫺ T)L, where k is the coefficient of thermal expansion,
Tƒ the tape’s field temperature, T the calibrated temperature of the
tape, and L the measured length. Determine the error with respect
to Tƒ.
(c) CP ⫽ (Pƒ ⫺ P)L/AE, where Pƒ is the field tension, P the tension
calibrated for the tape, A its cross-sectional area, E the modulus of
elasticity, and L the measured length. Determine the error with
respect to Pƒ.
(d) CS ⫽ ⫺w2l3s / 24P2ƒ, where w is the weight per unit length of the
tape, ls the length between supports, and Pƒ the field tension. Determine the error with respect to Pƒ.
6.12 Compute the corrected distance and its expected error if the measured
distance is 145.67 ft. Assume that Tƒ ⫽ 45 Ⳳ 5⬚F, Pƒ ⫽ 16 Ⳳ 1 lb,
there was a reading error of Ⳳ0.01 ft, and that the distance was mea-
PROBLEMS
97
sured as two end-support distances of 100.00 ft and 86.87 ft.
(Reminder: Do not forget the correction for length: CL ⫽ [(l ⫺ l⬘)/
l⬘]L, where l is the actual tape length, l⬘ its nominal length, and L the
measured line length.) The tape calibration data are given as follows.
A ⫽ 0.004 in2
l ⫽ 100.012 ft
w ⫽ 0.015 lb
l⬘ ⫽ 100 ft
k ⫽ 0.00000645⬚F
E ⫽ 29,000,000 lb/in2
P ⫽ 10 lb
T ⫽ 68⬚F
6.13 Show that Equation (6.12) is equivalent to Equation (6.11) for linear
equations.
6.14 Derive an expression similar to Equation (6.9) for the function z ⫽
a1x1 ⫹ a2x2 ⫹ a3x3.
6.15 The elevation of point C on the chimney shown in Figure 6.3 is desired.
Field angles and distances are observed. Station A has an elevation of
345.618 Ⳳ 0.008 m and station B has an elevation of 347.758 Ⳳ 0.008
m. The instrument height, hiA, at station A is 1.249 Ⳳ 0.003 m, and
the instrument height, hiB, at station B is 1.155 Ⳳ 0.003 m. Zenith
angles are read in the field. The other observations and their errors are
AB ⫽ 93.505 Ⳳ 0.006 m
A ⫽ 44⬚12⬘34ⴖ Ⳳ 7.9ⴖ
B ⫽ 39⬚26⬘56ⴖ Ⳳ 9.8ⴖ
z1 ⫽ 81⬚41⬘06ⴖ Ⳳ 12.3ⴖ
z2 ⫽ 84⬚10⬘25ⴖ Ⳳ 11.6ⴖ
What are the elevation of the chimney and the standard deviation in this
elevation?
Practical Exercises
6.16 With an engineer’s scale, measure the radius of the circle in the Figure
P6.16 ten times using different starting locations on the scale. Use a
magnifying glass and interpolate the readings on the scale to a tenth
of the smallest graduated reading on the scale.
(a) What are the mean radius of the circle and its standard deviation?
(b) Compute the area of the circle and its standard deviation.
98
PROPAGATION OF RANDOM ERRORS IN INDIRECTLY MEASURED QUANTITIES
(c) Calibrate a planimeter by measuring a 2-in. square. Calculate the
mean constant for the planimeter (k ⫽ units/4 in2), and based on
10 measurements, determine the standard deviation in the constant.
(d) Using the same planimeter, measure the area of the circle and
determine its standard deviation.
Figure P6.16
6.17 Develop a computational worksheet that solves Problem 6.11.
CHAPTER 7
ERROR PROPAGATION IN ANGLE
AND DISTANCE OBSERVATIONS
7.1
INTRODUCTION
All surveying observations are subject to errors from varying sources. For
example, when observing an angle, the major error sources include instrument
placement and leveling, target placement, circle reading, and target pointing.
Although great care may be taken in observing the angle, these error sources
will render inexact results. To appreciate fully the need for adjustments, surveyors must be able to identify the major observational error sources, know
their effects on the measurements, and understand how they can be modeled.
In this chapter, emphasis is placed on analyzing the errors in observed horizontal angles and distances. In Chapter 8 the manner in which these errors
propagate to produce traverse misclosures is studied. In Chapter 9 the propagation of angular errors in elevation determination is covered.
7.2
ERROR SOURCES IN HORIZONTAL ANGLES
Whether a transit, theodolite, or total station instrument is used, errors are
present in every horizontal angle observation. Whenever an instrument’s circles are read, a small error is introduced into the final angle. Also, in pointing
to a target, a small amount of error always occurs. Other major error sources
in angle observations include instrument and target setup errors and the instrument leveling error. Each of these sources produces random errors. They
may be small or large, depending on the instrument, the operator, and the
conditions at the time of the angle observation. The effects of reading, pointAdjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
99
100
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
ing, and leveling errors can be reduced by increasing the number of angle
repetitions. However, the effects of instrument and target setup errors can be
reduced only by increasing sight distances.
7.3
READING ERRORS
Errors in reading conventional transits and theodolites depend on the quality
of the instrument’s optics, the size of the smallest division of the circle, and
the operator’s abilities: for example, the ability to set and read a transit vernier, or to set and read the micrometer of a theodolite. Typical reading errors
for a 1ⴖ micrometer theodolite can range from tenths of a second to several
seconds. Reading errors also occur with digital instruments, their size being
dependent on the sensitivity of the particular electronic angular resolution
system. Manufacturers quote the estimated combined pointing and reading
precision for an individual direction measured face I (direct) and face II (reversed) with their instruments in terms of standard deviations. Typical values
range from 1ⴖ for the more precise instruments to 10ⴖ for the less expensive ones. These errors are random, and their effects on an angle depend
on the observation method and the number of repeated observations.
7.3.1
Angles Observed by the Repetition Method
When observing a horizontal angle by repetition using a repeating instrument,
the circle is first zeroed so that angles can be accumulated on the horizontal
circle. The angle is turned a number of times, and finally, the cumulative
angle is read and divided by the number of repetitions, to determine the
average angular value. In this method, a reading error exists in just two positions, regardless of the number of repetitions. The first reading error occurs
when the circle is zeroed and the second when reading the final cumulative
angle. For this procedure, the average angle is computed as
␣⫽
␣1 ⫹ ␣2 ⫹ ⫹ ␣n
n
(a)
where ␣ is the average angle, and ␣1, ␣2, . . . , ␣n are the n repetitions of the
angle. Recognizing that readings occur only when zeroing the plates and
reading the final direction ␣n and applying Equation (6.16) to Equation (a),
the standard error in reading the angle using the repetition method is
␣r ⫽
20 ⫹ r2
n
(7.1)
7.3
READING ERRORS
101
where ␣r is the error in the average angle due to reading, 0 the estimated
error in setting zero on the circle, r the estimated error in the final reading,
and n the number of repetitions of the angle. Note that the number of repetitions should always be an even number, with half being turned face I (direct)
and half face II (reversed). This procedure compensates for systematic instrumental errors.
Assuming that the observer’s ability to set zero and to read the circle are
equal, Equation (7.1) is simplified to
␣r ⫽
r2
n
(7.2)
Example 7.1 Suppose that an angle is turned six times using the repetition
method. For an observer having a personal reading error of 1.5ⴖ, what is
the error in the final angle due to circle reading?
SOLUTION From Equation (7.2),
1.5ⴖ2
⫽ 0.4ⴖ
6
␣r ⫽
7.3.2
Angles Observed by the Directional Method
When a horizontal angle is observed by the directional method, the horizontal
circle is read in both the backsight and foresight directions. The angle is then
the difference between the two readings. Multiple observations of the angle
are made, with the circle being advanced prior to each reading to compensate
for the systematic errors. The final angle is taken as the average of all values
observed. Again, an even number of repetitions are made, with half taken in
the face I and half in the face II position. Since each repetition of the angle
requires two readings, the error in the average angle due to the reading error
is computed using Equation (6.16), which yields
␣r ⫽
(2r1b ⫹ 2r1ƒ) ⫹ (2r2b ⫹ 2r2ƒ) ⫹ ⫹ (2rnb ⫹ 2rnƒ)
n
(7.3)
where rib and riƒ are the estimated errors in reading the circle for the backsight and foresight directions, respectively, and n is the number of repetitions.
Assuming that one’s ability to read the circle is independent of the particular
direction, so that rib ⫽ riƒ ⫽ r, Equation (7.3) simplifies to
102
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
␣r ⫽
r2
n
(7.4)
Example 7.2 Using the same parameters of six repetitions and an estimated
observer reading error of 1.5ⴖ as given in Example 7.1, find the error estimated in the average angle due to reading when the directional method is
used.
SOLUTION
␣P ⫽
1.5ⴖ2
6
⫽ 0.9ⴖ
Note that the additional readings required in the directional method produce
a larger error in the angle than that obtained using the repetition method.
7.4
POINTING ERRORS
Accuracy in pointing to a target depends on several factors. These include
the optical qualities of the instrument, target size, the observer’s personal
ability to place the crosswires on a target, and the weather conditions at the
time of observation. Pointing errors are random, and they will occur in every
angle observation no matter the method used. Since each repetition of an
angle consists of two pointings, the pointing error for an angle that is the
mean of n repetitions can be estimated using Equation (6.16) as
␣p ⫽
22p1 ⫹ 22p2 ⫹ ⫹ 22pn
n
(7.5)
where ␣p is the error due to pointing and p1, p2, . . . , pn are the estimated
errors in pointings for the first repetition, second repetition, and so on. Again
for a given instrument and observer, the pointing error can be assumed the
same for each repetition (i.e., p1 ⫽ p2 ⫽ ⫽ pn ⫽ p), and Equation
(7.5) simplifies to
␣p ⫽
p2
n
(7.6)
Example 7.3 An angle is observed six times by an observer whose ability
to point on a well-defined target is estimated to be 1.8ⴖ. What is the estimated error in the average angle due to the pointing error?
7.5
ESTIMATED POINTING AND READING ERRORS WITH TOTAL STATIONS
103
SOLUTION From Equation (7.6),
1.8ⴖ2
␣p ⫽
6
⫽ 1.0ⴖ
7.5 ESTIMATED POINTING AND READING ERRORS WITH
TOTAL STATIONS
With the introduction of electronic theodolites and subsequently, total station
instruments, new standards were developed for estimating errors in angle
observations. The new standards, called DIN 18723, provide values for estimated errors in the mean of two direction observations, one each in the face
I and face II positions. Thus, in terms of a single pointing and reading error,
pr, the DIN value, DIN, can be expressed as
DIN ⫽
pr2
2
⫽
pr
2
Using this equation, the expression for the estimated error in the observation
of a single direction due to pointing and reading with an electronic theodolite
is
pr ⫽ DIN 2
(b)
Using a procedure similar to that given in Equation (7.6), the estimated error
in an angle measured n times and averaged due to pointing and reading is
␣pr ⫽
pr2
n
(c)
Substituting Equation (b) into Equation (c) yields
␣pr ⫽
2DIN
n
(7.7)
Example 7.4 An angle is observed six times by an operator with a total
station instrument having a published DIN 18723 value for the pointing and
reading error of 5ⴖ. What is the estimated error in the angle due to the
pointing and reading error?
104
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
SOLUTION From Equation (7.7),
␣pr ⫽
7.6
2 ⫻ 5ⴖ
⫽ 4.1ⴖ
6
TARGET CENTERING ERRORS
Whenever a target is set over a station, there will be some error due to faulty
centering. This can be attributed to environmental conditions, optical plummet
errors, quality of the optics, plumb bob centering error, personal abilities, and
so on. When care is taken, the instrument is usually within 0.001 to 0.01 ft
of the true station location. Although these sources produce a constant centering error for any particular angle, it will appear as random in the adjustment
of a network involving many stations since targets and instruments will center
differently over a point. This error will also be noticed in resurveys of the
same points.
An estimate of the effect of this error in an angle observation can be made
by analyzing its contribution to a single direction. As shown in Figure 7.1,
the angular error due to the centering error depends on the position of the
target. If the target is on line but off center, as shown in Figure 7.1(a), the
target centering error does not contribute to the angular error. However, as
the target moves to either side of the sight line, the error size increases. As
shown in Figure 7.1(d), the largest error occurs when the target is offset
perpendicular to the line of sight. Letting d represent the distance the target
Figure 7.1 Possible target locations.
7.6
TARGET CENTERING ERRORS
105
is from the true station location, from Figure 7.1(d), the maximum error in
an individual direction due to the target centering error is
e⫽
d
rad
D
(7.8)
where e is the uncertainty in the direction due to the target centering error,
d the amount of a centering error at the time of pointing, and as shown in
Figure 7.2, D is the distance from the instrument center to the target.
Since two directions are required for each angle observation, the contribution of the target centering error to the total angular error is
␣t ⫽
冪冉 冊 冉 冊
d1
D1
2
⫹
d2
2
D2
(7.9)
where ␣t is the angular error due to the target centering error, d1 and d2
are the target centering errors at stations 1 and 2, respectively, and D1 and
D2 are the distances from the target to the instrument at stations 1 and 2,
respectively. Assuming the ability to center the target over a point is independent of the particular direction, it can be stated that d1 ⫽ d2 ⫽ t. Finally,
the results of Equation (7.9) are unitless. To convert the result to arc seconds,
it must be multiplied by the constant (206,264.8ⴖ /rad), which yields
兹D21 ⫹ D22
ⴖ␣t ⫽
D1D2
t
(7.10)
Notice that the same target centering error occurs on each pointing. Thus,
it cannot be reduced in size by taking multiple pointings, and therefore Equation (7.10) is not divided by the number of angle repetitions. This makes the
target centering error one of the more significant errors in angle observations.
Figure 7.2 Error in an angle due to target centering error.
106
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
It also shows that the only method to decrease the size of this error is to
increase the sight distances.
Example 7.5 An observer’s estimated ability at centering targets over a station is 0.003 ft. For a particular angle observation, the backsight and foresight distances from the instrument station to the targets are approximately
250 ft and 450 ft, respectively. What is the angular error due to the error in
target centering?
SOLUTION From Equation (7.10), the estimated error is
2502 ⫹ 4502
ⴖ␣t ⫽
250 ⫻ 450
0.003 ⫻ 206,264.8ⴖ /rad ⫽ 2.8ⴖ
If handheld range poles were used in this example with an estimated centering
error of 0.01 ft, the estimated angular error due to the target centering would
be
2502 ⫹ 4502
ⴖ␣t ⫽
250 ⫻ 450
0.01 ⫻ 206,264.8ⴖ /rad ⫽ 9.4ⴖ
Obviously, this is a significant error source if care is not taken in target
centering.
7.7
INSTRUMENT CENTERING ERRORS
Every time an instrument is centered over a point, there is some error in its
position with respect to the true station location. This error is dependent on
the quality of the instrument and the state of adjustment of its optical plummet, the quality of the tripod, and the skill of surveyor. The error can be
compensating, as shown in Figure 7.3(a), or it can be maximum when the
instrument is on the angle bisector, as shown in Figure 7.3(b) and (c). For
any individual setup, this error is a constant; however, since the instrument’s
location is random with respect to the true station location, it will appear to
be random in the adjustment of a network involving many stations. Like the
target centering error, it will appear also during a resurvey of the points. From
Figure 7.3, the true angle ␣ is
␣ ⫽ (P2 ⫹ ε2) ⫺ (P1 ⫹ ε1) ⫽ (P2 ⫺ P1) ⫹ (ε2 ⫺ ε1)
where P1 and P2 are the true directions and ε1 and ε2 are errors in those
directions due to faulty instrument centering. The error size for any setup is
7.7
INSTRUMENT CENTERING ERRORS
107
Figure 7.3 Error in angle due to error in instrument centering.
ε ⫽ ε2 ⫺ ε1
(7.11)
The error in the observed angle due to instrument centering errors is analyzed by propagating errors in a formula based on (x,y) coordinates. In Figure 7.4 a coordinate system has been constructed with the x axis going from
the true station to the foresight station. The y axis passes through the instrument’s vertical axis and is perpendicular to the x axis. From the figure the
following equations can be derived:
ih ⫽ ip ⫺ qr
(7.12)
ih ⫽ iq cos ␣ ⫺ sq sin ␣
Letting sq ⫽ x and iq ⫽ y, Equation (7.12) can be rewritten as
ih ⫽ y cos ␣ ⫺ x sin ␣
(7.13)
Furthermore, in Figure 7.4,
ε1 ⫽
ih
y cos ␣ ⫺ x sin ␣
⫽
D1
D1
(7.14)
ε2 ⫽
y
D2
(7.15)
By substituting Equations (7.14) and (7.15) into Equation (7.11), the error
in an observed angle due to the instrument centering error is
ε⫽
y
y cos ␣ ⫺ x sin ␣
⫺
D2
D1
Reorganizing Equation (7.16) yields
(7.16)
108
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
Figure 7.4 Analysis of instrument centering error.
ε⫽
D1y ⫹ D2 x sin ␣ ⫺ D2 y cos ␣
D1D2
(7.17)
Now because the instrument’s position is truly random, Equation (6.16)
can be used to find the angular uncertainty due to the instrument centering
error. Taking the partial derivative of Equation (7.17) with respect to both x
and y gives
⭸ε
D sin ␣
⫽ 2
⭸x
D1D2
(7.18)
⭸ε
D ⫺ D2 cos ␣
⫽ 1
⭸y
D1D2
Now substituting the partial derivatives in Equation (7.18) into Equation
(6.16) gives
ε2 ⫽
D2 sin ␣ 2 D1 ⫺ D2 cos ␣ 2
x ⫹
y
D1D2
D1D2
(7.19)
7.7
INSTRUMENT CENTERING ERRORS
109
Figure 7.5 Instrument centering errors at a station.
Because this error is a constant for a setup, the mean angle has the same
error as a single angle, and thus it is not reduced by taking several repetitions.
The estimated error in the position of a station is derived from a bivariate
distribution,1 where the coordinate components are independent and have
equal magnitudes. Assuming that estimated errors in the x and y axes are x
and y, from Figure 7.5 it is seen that
x ⫽ y ⫽
i
2
Letting ε ⫽ ␣i, expanding the squares of Equation (7.19), and rearranging
yields
␣i ⫽
D21 ⫹ D22 (cos2␣ ⫹ sin2␣) ⫺ 2D1D2 cos ␣ 2i
D21D22
2
(7.20)
Making the trigonometric substitutions of cos2␣ ⫹ sin2␣ ⫽ 1 and D21 ⫹
D ⫺ 2D1D2 cos ␣ ⫽ D23 in Equation (7.20), taking the square root of both
sides, and multiplying by (206,264.8ⴖ /rad) to convert the results to arc
seconds yields
2
2
D3 i
D1D2 2
ⴖ␣i ⫽
(7.21)
Example 7.6 An observer centers the instrument to within 0.005 ft of a
station for an angle with backsight and foresight distances of 250 ft and 450
ft, respectively. The angle observed is 50⬚. What is the error in the angle due
to the instrument centering error?
1
The bivariate distribution is discussed in Chapter 19.
110
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
SOLUTION Using the cosine law, D23 ⫽ D21 ⫹ D22 ⫺ 2D1D2 cos ∠, and
substituting in the appropriate values, we find D3 to be
D3 ⫽ 2502 ⫹ 4502 ⫺ 2 ⫻ 250 ⫻ 450 ⫻ cos 50⬚ ⫽ 346.95 ft
Substituting this value into Equation (7.21), the estimated contribution of the
instrument centering error to the overall angular error is
ⴖ␣i ⫽
7.8
346.95 0.005
206,264.8ⴖ /rad ⫽ 2.2ⴖ
250 ⫻ 450 2
EFFECTS OF LEVELING ERRORS IN ANGLE OBSERVATIONS
If an instrument is imperfectly leveled, its vertical axis is not truly vertical
and its horizontal circle and horizontal axis are both inclined. If while an
instrument is imperfectly leveled it is used to measure horizontal angles, the
angles will be observed in a plane other than horizontal. Errors that result
from this error source are most severe when the backsights and foresights are
steeply inclined: for example, in making astronomical observations or traversing over mountains. If the bubble of a theodolite were to remain off center
by the same amount during the entire angle-observation process, the resulting
error would be systematic. However, because an operator normally monitors
the bubble carefully and attempts to keep it centered while turning angles,
the amount and direction by which the instrument is out of level becomes
random, and hence the resulting errors tend to be random. Even if the operator
does not monitor the instrument’s level, this error will appear to be random
in a resurvey.
In Figure 7.6, ε represents the angular error that occurs in either the backsight or foresight of a horizontal angle observation made with an instrument
out of level and located at station I. The line of sight IS is elevated by the
vertical angle v. In the figure, IS is shown perpendicular to the instrument’s
horizontal axis. The amount by which the instrument is out of level is ƒd,
where ƒd is the number of fractional divisions the bubble is off center and
is the sensitivity of the bubble. From the figure,
SP ⫽ D tan v
(d)
PP⬘ ⫽ Dε
(e)
and
7.8
EFFECTS OF LEVELING ERRORS IN ANGLE OBSERVATIONS
111
Figure 7.6 Effects of instrument leveling error.
where D is the horizontal component of the sighting distance and the angular
error ε is in radians. Because the amount of leveling error is small, PP⬘ can
be approximated as a circular arc, and thus
PP⬘ ⫽ ƒd(SP)
(f)
Substituting Equation (d) into Equation (f) yields
PP⬘ ⫽ ƒdD tan v
(7.22)
Now substituting Equation (7.22) into Equation (e) and reducing, the error in
an individual pointing due to the instrument leveling error is
ε ⫽ ƒd tan v
(7.23)
As noted above, Figure 7.6 shows the line of sight oriented perpendicular
to the instrument’s horizontal axis. Also, the direction in which a bubble runs
is random. Thus, Equation (6.18) can be used to compute the combined angular error that results from n repetitions of an angle made with an imperfectly
leveled instrument (note that each angle measurement involves both backsight
and foresight pointings):
(ƒd tan vb)2 ⫹ (ƒd tan vƒ)2
␣l ⫽
n
(7.24)
where vb and vƒ are the vertical angles to the backsight and foresight targets,
respectively, and n is the number of repetitions of the angle.
112
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
Example 7.7 A horizontal angle is observed on a mountainside where the
backsight is to the peak and the foresight is in the valley. The average zenith
angles to the backsight and foresight are 80⬚ and 95⬚, respectively. The instrument has a level bubble with a sensitivity of 30ⴖ /div and is leveled to
within 0.3 div. For the average angle obtained from six repetitions, what is
the contribution of the leveling error to the overall angular error?
SOLUTION The zenith angles converted to vertical angles are ⫹10⬚ and
⫺5⬚, respectively. Substituting the appropriate values into Equation (7.24)
yields
ⴖ /div) tan 10⬚]
冪[0.3⫹ div(30
[0.3 div(30ⴖ /div) tan(⫺5⬚)]
2
2
␣l ⫽
6
⫽ 0.7ⴖ
This error is generally small for traditional surveying work when normal
care is taken in leveling the instrument. Thus, it can generally be ignored for
all but the most precise work. However, as noted earlier, for astronomical
observations this error can become quite large, due to the steeply inclined
sights to celestial objects. Thus, for astronomical observations it is extremely
important to keep the instrument leveled precisely for each observation.
7.9 NUMERICAL EXAMPLE OF COMBINED ERROR PROPAGATION
IN A SINGLE HORIZONTAL ANGLE
Example 7.8 Assume that an angle is observed four times with a directionaltype instrument. The observer has an estimated reading error of 1ⴖ and a
pointing error of 1.5ⴖ. The targets are well defined and placed on an optical
plummet tribrach with an estimated centering error of 0.003 ft. The instrument is in adjustment and centered over the station to within 0.003 ft. The
horizontal distances from the instrument to the backsight and foresight targets
are approximately 251 ft and 347 ft, respectively. The average angle is
65⬚37⬘12ⴖ. What is the estimated error in the angle observation?
SOLUTION The best way to solve this type of problem is to computed
estimated errors for each item in Sections 7.3 to 7.8 individually, and then
apply Equation (6.18).
Error due to reading: Substituting the appropriate values into Equation
(7.4) yields
7.9
NUMERICAL EXAMPLE OF COMBINED ERROR PROPAGATION
1ⴖ2
␣r ⫽
4
113
⫽ 0.71ⴖ
Error due to pointing: Substituting the appropriate values into Equation
(7.6) yields
␣p ⫽
1.5ⴖ2
4
⫽ 1.06ⴖ
Error due to target centering: Substituting the appropriate values into
Equation (7.10) yields
ⴖ␣t ⫽
2512 ⫹ 3472
251 ⫻ 347
(0.003)206,264.8ⴖ /rad ⫽ 3.04ⴖ
Error due to instrument centering: From the cosine law we have
D23 ⫽ 2512 ⫹ 3472 ⫺ 2(251)(347) cos (65⬚37⬘12ⴖ)
D3 ⫽ 334 ft
Substituting the appropriate values into Equation (7.21) yields
334
0.003
206,264.8ⴖ /rad ⫽ 1.68ⴖ
251 ⫻ 347 2
ⴖ␣t ⫽
Combined error: From Equation (6.18), the estimated angular error is
␣ ⫽ 0.712 ⫹ 1.062 ⫹ 3.042 ⫹ 1.682 ⫽ 3.7ⴖ
In Example 7.8, the largest error sources are due to the target and instrument centering errors, respectively. This is true even when the estimated error
in centering the target and instrument are only 0.003 ft. Unfortunately,
many surveyors place more confidence in their observations than is warranted.
Since these two error sources do not decrease with increased repetitions, there
is a limit to what can be expected from any survey. For instance, assume that
the targets were handheld reflector poles with an estimated centering error of
0.01 ft. Then the error due to the target centering error becomes 10.1ⴖ.
This results in an estimated angular error of 10.3ⴖ. If a 99% probable error
were computed, a value as large as 60ⴖ would be possible!
114
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
7.10 USE OF ESTIMATED ERRORS TO CHECK ANGULAR
MISCLOSURE IN A TRAVERSE
When a traverse is closed geometrically, the angles are generally checked for
misclosure. By computing the errors for each angle in the traverse as described in Section 7.9 and summing the results with Equation (6.18), an estimate for the size of the angular misclosure is obtained. The procedure is
best demonstrated with an example.
Example 7.9 Assume that each of the angles in Figure 7.7 was observed
using four repetitions (twice direct and twice reverse) and their estimated
errors were computed as shown in Table 7.1. Does this traverse meet acceptable angular closure at a 95% level of confidence?
SOLUTION The actual angular misclosure of the traverse is 30ⴖ. The estimated angular misclosure of the traverse is found by applying Equation
(6.18) with the errors computed for each angle. That is, the estimated angular
misclosure is
兺∠ ⫽ 8.92 ⫹ 12.12 ⫹ 13.72 ⫹ 10.02 ⫹ 9.92 ⫽ 24.7ⴖ
Thus, the actual angular misclosure of 30ⴖ is greater than the value estimated
(24.7ⴖ) at a 68.3% probable error level. However, since each angle was turned
only four times, a 95% probable error must be computed by using the appropriate t value from Table D.3.
This problem begs the question of what the appropriate number of degrees
of freedom is for the summation of the angles. Only four of the angles are
required in the summation since the fifth angle can be computed from the
other four and thus is redundant. Since each angle is turned four times, it can
be argued that there are 16 redundant observations: that is, 12 angles at the
first four stations and four at the fifth station. However, this assumes that
instrumental systematic errors were not present in the observational process
since only the average of a face I and a face II reading can eliminate system-
Figure 7.7 Close polygon traverse.
7.10
USE OF ESTIMATED ERRORS TO CHECK ANGULAR MISCLOSURE IN A TRAVERSE
115
TABLE 7.1 Data for Example 7.9
Angle
Observed
Value
Computed
Angular Error
1
2
3
4
5
60⬚50⬘48ⴖ
134⬚09⬘24ⴖ
109⬚00⬘12ⴖ
100⬚59⬘54ⴖ
135⬚00⬘12ⴖ
8.9ⴖ
12.1ⴖ
13.7ⴖ
10.0ⴖ
9.9ⴖ
atic errors in the instrument. If there are n angles and each angle is turned r
times, the total number of redundant observations would be n(r ⫺ 1) ⫹ 1. In
this case it would be 5(4 ⫺ 1) ⫹ 1 ⫽ 16.
A second approach is to account for instrumental systematic errors when
counting redundant observations. This method requires that an angle exists
only if it is observed with both faces of the instrument. In this case there is
one redundant angle at each of the first four stations, with the fifth angle
having two redundant observations, for a total of six redundant observations.
Using this argument, the number of redundant angles in the traverse would
be n(r/2 ⫺ 1) ⫹ 1. In this example it would be 5(4/2 ⫺ 1) ⫹ 1 ⫽ 6.
A third approach would be to consider each mean angle observed at each
station to be a single observation, since only mean observations are being
used in the computations. In this case there would be only one redundant
angle for the traverse. However, had horizon closures been observed at each
station, the additional angles would add n redundant observations.
A fourth approach would be to determine the 95% probable error at each
station and then use Equation (6.18) to sum these 95% error values. In this
example, each station has three redundant observations. In general, there
would be r ⫺ 1 redundant angle observations, where r represents the number
of times that the angle was repeated during the observation process.
The last two methods are the most conservative since they allow the most
error in the sum of the angles. The fourth method is used in this book. However, a surveyor must decide which method is most appropriate for his or her
practice. As stated in Chapter 5, the statistician must make decisions when
performing any test. Using the fourth method, there are three redundant observation at each station. To finish the problem, we construct a 95% confidence interval, or perform a two-tailed test to determine the range of error
that is statistically equal to zero. In this case, t0.025,3 ⫽ 3.183, the 95% probable
error for the angular sum is
95% ⫽ 3.183 ⫻ 24.7ⴖ ⫽ 78.6ⴖ
Thus, the traverse angles are well within the range of allowable error. We
cannot reject the null hypothesis that the error in the angles is not statistically
116
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
equal to zero. Thus, the survey meets the minimum level of angular closure
at a 95% probable error. However, it must be remembered that because of
the possibility of Type II errors, we can only state that there is no statistical
reason to believe that there is a blunder in the angle observations.
Example 7.9 presents another question for the statistician or surveyor. That
is, should a surveyor allow a field crew to have this large an angular misclosure in the traverse? Statistically, the answer would seem to be yes, but it
must remembered that the target and instrument centering errors affect angle
observations only if the instrument and targets are reset after each observation.
Since this is never done in practice, these two errors should not be included
in the summation of the angles. Instead, the allowable angular misclosure
should be based solely on pointing and reading errors. For example, if the
angles were observed with a total station having a DIN 18723 standard of
1ⴖ, by Equation (7.7) the pointing and reading error for each angle would
be
2 ⫻ 1ⴖ
⫽ 1.4ⴖ
2
␣pr ⫽
By Equation (6.19), the error in the summation of the five angles would be
1.4ⴖ5 ⫽ 3.2ⴖ. Using the same critical t value of 3.183, the allowable
error in the angular misclosure should only be 3.183 ⫻ 3.2ⴖ ⫽ 10ⴖ. If this
instrument had be used in Example 7.9, the field-observed angular closure of
30ⴖ would be unacceptable and would warrant reobservation of some or all
of the angles.
As stated in Sections 7.6 and 7.7, the angular misclosure of 78.6ⴖ computed
in Example 7.9 will be noticed only when the target and instrument are reset
on a survey. This will happen during the resurvey, when the centering errors
of the target and instrument from the original survey will be present in the
record directions. Thus, record azimuths and/or bearings could disagree from
those determined in the resurvey by this amount, assuming that the equipment
used in the resurvey is comparable or of higher quality than that used in the
original survey.
7.11 ERRORS IN ASTRONOMICAL OBSERVATIONS FOR
AN AZIMUTH
The total error in an azimuth determined from astronomical observations depends on errors from several sources, including those in timing, the observer’s
latitude and longitude, the celestial object’s position at observation time, timing accuracy, observer response time, instrument optics, atmospheric condi-
7.11
ERRORS IN ASTRONOMICAL OBSERVATIONS FOR AN AZIMUTH
117
tions, and others, as identified in Section 7.2. The error in an astronomical
observation can be estimated by analyzing the hour–angle formula, which is
sin t
cos tan ␦ ⫺ sin cos t
z ⫽ tan⫺1
(7.25)
In Equation (7.25), z is the azimuth of the celestial object at the time of the
observation, t the t angle of the PZS triangle at the time of observation, the
observer’s latitude, and ␦ the object’s declination at the time of the
observation.
The t angle is a function of the local hour angle (LHA) of the sun or a
star at the time of observation. That is, when the LHA ⬍ 180⬚, t ⫽ LHA;
otherwise, t ⫽ 360⬚ ⫺ LHA. Furthermore, LHA is a function of the Greenwich
hour angle (GHA) of the celestial body and the observer’s longitude; that is,
LHA ⫽ GHA ⫹
(7.26)
where is the observer’s longitude, considered positive for eastern longitude
and negative for western longitude. The GHA increases approximately 15⬚
per hour of time, and thus an estimate of the error in the GHA is
approximately
t ⫽ 15⬚ ⫻ T
where T is the estimated error in time (in hours). Similarly, by using the
declination at 0h and 24h, the amount of change in declination per second can
be derived and thus the estimated error in declination determined.
Using Equation (6.16), the error in a star’s azimuth is estimated by taking
the partial derivative of Equation (7.25) with respect to t, ␦, , and . To do
this, simplify Equation (7.25) by letting
F ⫽ cos tan ␦ ⫺ sin cos t
(7.27a)
u ⫽ sin t ⫻ F⫺1
(7.27b)
and
Substituting in Equations (7.27), Equation (7.25) is rewritten as
z ⫽ tan⫺1
From calculus it is known that
sin t
⫽ tan⫺1u
F
(7.28)
118
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
d tan⫺1u
1 du
⫽
dx
1 ⫹ u2 dx
Applying this fundamental relation to Equation (7.28) and letting G represent
GHA yields
⭸z
1
du
F2
du
⫽
⫽
2
2
2
⭸G
1 ⫹ [sin(G ⫺ )/F] dG F ⫹ sin (G ⫺ ) dG
(7.29)
Now du/dG is
du
cos(G ⫺ ) sin(G ⫺ )
⫽
⫺
sin sin(G ⫺ )
dG
F
F2
⫽
cos(G ⫺ ) sin2(G ⫺ ) sin
⫺
F
F2
and thus,
du
F cos(G ⫺ ) ⫺ sin2(G ⫺ ) sin
⫽
dG
F2
(7.30)
Substituting Equation (7.30) into Equation (7.29) and substituting in t for G
⫺ yields
⭸z
F cos t ⫺ sin2t sin
⫽
⭸G
F2 ⫹ sin2t
(7.31)
In a similar fashion, the following partial derivatives are developed from
Equation (7.25):
dz
sin t cos
⫽⫺ 2
d␦
cos ␦(F2 ⫹ sin2t)
(7.32)
⭸z
sin t cos t cos ⫹ sin t sin tan ␦
⫽
⭸
F2 ⫹ sin2t
(7.33)
⭸z
sin2t sin ⫺ F cost t
⫽
⭸
F2 ⫹ sin2t
(7.34)
where t is the t angle of the PZS triangle, z the celestial object’s azimuth, ␦
the celestial object’s declination, the observer’s latitude, the observer’s
longitude, and F ⫽ cos tan ␦ ⫺ sin cos t.
7.11
ERRORS IN ASTRONOMICAL OBSERVATIONS FOR AN AZIMUTH
119
If the horizontal angle, H, is the angle to the right observed from the line
to the celestial body, the equation for a line’s azimuth is
Az ⫽ z ⫹ 360⬚ ⫺ H
Therefore, the error contributions from the horizontal angle observation must
be included in computing the overall error in the azimuth. Since the distance
to the star is considered infinite, the estimated contribution to the angular
error due to the instrument centering error can be determined with a formula
similar to that for the target centering error with one pointing. That is,
␣i ⫽
i
D
(7.35)
where i is the centering error in the instrument and D is the length of the
azimuth line in the same units. Note that the results of Equation (7.35) are
in radian units and must be multiplied by to yield a value in arc seconds.
Example 7.10 Using Equation (7.25), the azimuth to Polaris was found to
be 0⬚01⬘31.9ⴖ. The observation time was 1:00:00 UTC with an estimated error
of T ⫽ 0.5s. The Greenwich hour angles to the star at 0h and 24h UTC
were 243⬚27⬘05.0ⴖ and 244⬚25⬘50.0ⴖ, respectively. The LHA at the time of
the observation was 181⬚27⬘40.4ⴖ. The declinations at 0h and 24h were
89⬚13⬘38.18ⴖ and 89⬚13⬘38.16ⴖ, respectively. At the time of observation, the
declination was 89⬚13⬘38.18ⴖ. The clockwise horizontal angle measured from
the backsight to a target 450.00 ft was 221⬚25⬘55.9ⴖ. The observer’s latitude
and longitude were scaled from a map as 40⬚13⬘54ⴖN and 77⬚01⬘51.5ⴖW,
respectively, with estimated errors of 1ⴖ. The vertical angle to the star was
39⬚27⬘33.1ⴖ. The observer’s estimated errors in reading and pointing are 1ⴖ
and 1.5ⴖ, respectively, and the instrument was leveled to within 0.3 of a
division with a bubble sensitivity of 25ⴖ /div. The estimated error in instrument
and target centering is 0.003 ft. What are the azimuth of the line and its
estimated error? What is the error at the 95% level of confidence?
SOLUTION The azimuth of the line is Az ⫽ 0⬚01⬘31.9ⴖ ⫹ 360 ⫺
221⬚25⬘55.9ⴖ ⫽ 138⬚35⬘36ⴖ. Using the Greenwich hour angles at 0h and 24h,
an error of 0.5s time will result in an estimated error in the GHA of
360⬚ ⫹ (244⬚25⬘50.0ⴖ ⫺ 243⬚27⬘05.0ⴖ)
0.05s ⫽ 7.52ⴖ
24h ⫻ 3600s / h
Since t ⫽ 360⬚ ⫺ LHA ⫽ 178⬚32⬘19.6ⴖ, F in Equations (7.29) through (7.34)
is
120
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
F ⫽ cos(40⬚13⬘54ⴖ) tan(89⬚13⬘38.18ⴖ) ⫺ sin(40⬚13⬘54ⴖ) cos(178⬚32⬘19.6ⴖ)
⫽ 57.249
The error in the observed azimuth can be estimated by computing the individual error terms as follows:
(a) From Equation (7.31), the error with respect to the GHA, G, is
⭸z
⫽
⭸G G
57.249 cos(178⬚32⬘19.6ⴖ) ⫺ sin2(178⬚32⬘19.6ⴖ) sin(40⬚13⬘54ⴖ)
(7.52ⴖ)
57.2492 ⫹ sin2(178⬚32⬘19.6ⴖ)
⫽ 0.13ⴖ
(b) By observing the change in declination, it is obvious that for this observation, the error in a time of 0.5s is insignificant. In fact, for the
entire day, the declination changes only 0.02ⴖ. This situation is common for stars. However, the sun’s declination may change from only
a few seconds daily to more than 23 minutes per day, and thus for
solar observations, this error term should not be ignored.
(c) From Equation (7.33) the error with respect to latitude, , is
⭸z
sin t cos t cos ⫹ sin t sin tan ␦
⫽
⫽ 0.0004ⴖ
⭸
F2 ⫹ sin2t
(d) From Equation (7.34) the error with respect to longitude, , is
⭸z
⫽
⭸
sin2(178⬚32⬘19.6ⴖ) sin(40⬚13⬘54ⴖ) ⫺ 57.249 cos(178⬚32⬘19.6ⴖ)
(1ⴖ)
57.2492 ⫹ sin2(178⬚32⬘19.6ⴖ)
⫽ 0.02ⴖ
(e) The circles are read both when pointing on the star and on the azimuth
mark. Thus, from Equation (7.2), the reading contribution to the estimated error in the azimuth is
␣r ⫽ r2 ⫽ 1ⴖ2 ⫽ 1.41ⴖ
(f) Using Equation (7.6), the estimated error in the azimuth due to pointing
is
␣p ⫽ p2 ⫽ 1.5ⴖ2 ⫽ 2.12ⴖ
7.12
121
ERRORS IN ELECTRONIC DISTANCE OBSERVATIONS
(g) From Equation (7.8), the estimated error in the azimuth due to target
centering is
冉 冊
d
0.003
⫽
206,264.8ⴖ /rad ⫽ 1.37ⴖ
D
450
ⴖ␣t ⫽
(h) Using Equation (7.35), the estimated error in the azimuth due to instrument centering is
冉 冊
d
0.003
⫽
206,264.8ⴖ /rad ⫽ 1.37ⴖ
D
450
ⴖ␣i ⫽
(i) From Equation (7.23), the estimated error in the azimuth due to the
leveling error is
␣b ⫽ ƒd tan v ⫽ 0.3 ⫻ 25ⴖ tan(39⬚27⬘33.1ⴖ) ⫽ 6.17ⴖ
Parts (a) through (i) are the errors for each individual error source. Using
Equation (6.18), the estimated error in the azimuth observation is
) ⫹ (1.32ⴖ) ⫹ (0.02ⴖ) ⫹ (0.0004ⴖ)
冪(0.13⫹ ⴖ(2.12
ⴖ) ⫹ 2(1.37ⴖ) ⫹ (6.17ⴖ)
2
AZ ⫽
2
2
2
2
2
2
⫽ 7.0ⴖ
Using the appropriate t value of t0.025,1 from Table D.3, the 95% error is
Az ⫽ 12.705 ⫻ 7.0ⴖ ⫽ 88.9ⴖ
Notice that in this problem, the largest error source in the azimuth error is
caused by the instrument leveling error.
7.12
ERRORS IN ELECTRONIC DISTANCE OBSERVATIONS
All EDM observations are subject to instrumental errors that manufacturers
list as constant, a, and scalar, b, error. A typical specified accuracy is (a ⫹
b ppm). In this expression, a is generally in the range 1 to 10 mm, and b is
a scalar error that typically has the range 1 to 10 ppm. Other errors involved
in electronic distance observations stem from the target and instrument centering errors. Since in any survey involving several stations these errors tend
to be random, they should be combined using Equation (6.18). Thus, the
estimated error in an EDM observed distance is
D ⫽ 兹2i ⫹ 2t ⫹ a2 ⫹ (D ⫻ b ppm)2
(7.36)
122
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
where D is the error in the observed distance D, i the instrument centering
error, t the reflector centering error, and a and b the instrument’s specified
accuracy parameters.
Example 7.11 A distance of 453.87 ft is observed using an EDM with a
manufacturer’s specified accuracy of (5 mm ⫹ 10 ppm). The instrument is
centered over the station with an estimated error of 0.003 ft, and the reflector, which is mounted on a handheld prism pole, is centered with an
estimated error of 0.01 ft. What is the error in the distance? What is the
E95 value?
SOLUTION Converting millimeters to feet using the survey foot2 definition
gives us
0.005 m ⫻ 39.37 in./12 in. ⫽ 0.0164 ft
The scalar portion of the manufacturer’s estimated standard error is computed
as
distance ⫻ b
1,000,000
In this example, the error is 453.87 ⫻ 10/1,000,000 ⫽ 0.0045 ft. Thus, according to Equation (7.36), the distance error is
⫽ (0.003)2 ⫹ (0.01)2 ⫹ (0.164)2 ⫹ (0.0045)2 ⫽ 0.02 ft
Using the appropriate t value from Table 3.2, the 95% probable error is
E95 ⫽ 1.6449 ⫽ 0.03 ft
Notice in this example that the instrument’s constant error is the largest
single contributor to the overall error in the observed distance, and it is followed closely by the target centering error. Furthermore, since both errors are
constants, their contribution to the total error is unchanged regardless of the
distance. Thus, for this particular EDM instrument, distances under 200 ft
could probably be observed more accurately with a calibrated steel tape. However, this statement depends on the terrain and the skill of the surveyors in
using a steel tape.
2
The survey foot definition is 1 meter ⫽ 39.37 inches, exactly.
PROBLEMS
7.13
123
USE OF COMPUTATIONAL SOFTWARE
The computations demonstrated in this chapter are rather tedious and time
consuming when done by hand, and this often leads to mistakes. This problem, and many others in surveying that involve repeated computations of a
few equations with different values, can be done conveniently with a spreadsheet, worksheet, or program. On the CD that accompanies this book, the
electronic book prepared with Mathcad demonstrates the programming of the
computational examples in this chapter. When practicing the following problems, the reader should consider writing software to perform the aforementioned computations.
PROBLEMS
7.1
Plot a graph of vertical angles from 0⬚ to 50⬚ versus the error in horizontal angle measurement due to an instrument leveling error of 5ⴖ.
7.2
For a direction with sight distances to the target of 100, 200, 300, 400,
600, 1000, and 1500 ft, construct:
(a) a table of estimated standard deviations due to target centering
when d ⫽ 0.005 ft.
(b) a plot of distance versus the standard deviations computed in part
(a).
7.3
For an angle of size 125⬚ with equal sight distances to the target of
100, 200, 300, 400, 600, 1000, and 1500 ft, construct:
(a) a table of standard deviations due to instrument centering when i
⫽ 0.005 ft.
(b) a plot of distance versus the standard deviations computed in part
(a).
7.4
Assuming setup errors of i ⫽ 0.002 m and t ⫽ 0.005 m, what
is the estimated error in a distance of length 684.326 m using an EDM
with stated accuracies of 3 mm ⫹ 3 ppm?
7.5
Repeat Problem 7.4 for a distance of length 1304.597 m.
7.6
Assuming setup errors of i ⫽ 0.005 ft and t ⫽ 0.005 ft, what is
the estimated error in a distance of length 1234.08 ft using an EDM
with stated accuracies of 3 mm ⫹ 3 ppm?
7.7
Repeat Problem 7.6 for a target centering error of 0.02 ft and a distance
of length 423.15 ft.
7.8
A 67⬚13⬘46ⴖ angle having a backsight length of 312.654 m and a foresight length of 205.061 m is observed, with the total station twice
124
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
having a stated DIN 18723 accuracy of 3ⴖ. Assuming instrument and
target centering errors of i ⫽ 0.002 m and t ⫽ 0.005 m, what
is the estimated error in the angle?
7.9
Repeat Problem 7.8 for an angle of 107⬚07⬘39ⴖ observed four times, a
backsight length of 306.85 ft, a foresight length of 258.03 ft, instrument and target centering errors of i ⫽ 0.005 ft and t ⫽ 0.01
ft, respectively, and an instrument with a stated DIN 18723 accuracy
of 2ⴖ.
7.10 For the following traverse data, compute the estimated error for each
angle if DIN ⫽ 2ⴖ, i ⫽ 0.005 ft, t ⫽ 0.01 ft, and the angles
were each measured four times (twice direct and twice reverse). Does
the traverse meet acceptable angular closures at a 95% level of
confidence?
Station
Angle
Distance (ft)
A
B
C
D
A
62⬚33⬘11ⴖ
124⬚56⬘19ⴖ
60⬚44⬘08ⴖ
111⬚46⬘07ⴖ
221.85
346.55
260.66
349.17
7.11 A total station with a DIN 18723 value of 3ⴖ was used to turn the
angles in Problem 7.10. Do the problem assuming the same estimated
errors in instrument and target centering.
7.12 A total station with a DIN 18723 value of 5ⴖ was used to turn the
angles in Problem 7.10. Do the problem assuming the same estimated
errors in instrument and target centering.
7.13 For the following traverse data, compute the estimated error in each
angle if r ⫽ 3ⴖ, p ⫽ 2ⴖ, i ⫽ t ⫽ 0.005 ft, and the angles
were observed four times (twice direct and twice reverse) using the
repetition method. Does the traverse meet acceptable angular closures
at a 95% level of confidence?
Station
Angle
Distance (ft)
A
B
C
D
E
A
38⬚58⬘24ⴖ
148⬚53⬘30ⴖ
84⬚28⬘06ⴖ
114⬚40⬘24ⴖ
152⬚59⬘18ⴖ
321.31
276.57
100.30
306.83
255.48
125
PROBLEMS
7.14 A total station with a DIN 18723 value of 2ⴖ was used to turn the
angles in Problem 7.13. Repeat the problem for this instrument.
7.15 An EDM was used to measure the distances in Problem 7.10. The
manufacturer’s specified errors for the instrument are (3 mm ⫹ 3
ppm). Using i ⫽ 0.005 ft and r ⫽ 0.01 ft, calculate the error in
each distance.
7.16 Repeat Problem 7.15 for the distances in Problem 7.13. The manufacturer’s specified error for the instrument was (5 mm ⫹ 5 ppm). Use
i and r from Problem 7.13.
7.17 The following observations and calculations were made on a sun observation to determine the azimuth of a line:
Observation
No.
UTC
1
2
3
4
5
6
16:30:00
16:35:00
16:40:00
16:45:00
16:50:00
16:55:00
Horizontal Vertical
Angle
Angle
41⬚02⬘33ⴖ
42⬚35⬘28ⴖ
44⬚09⬘23ⴖ
45⬚44⬘25ⴖ
47⬚20⬘21ⴖ
48⬚57⬘24ⴖ
39⬚53⬘08ⴖ
40⬚16⬘49ⴖ
40⬚39⬘11ⴖ
41⬚00⬘05ⴖ
41⬚19⬘42ⴖ
41⬚37⬘47ⴖ
␦
LHA
z
⫺3⬚28⬘00.58ⴖ
⫺3⬚28⬘05.43ⴖ
⫺3⬚28⬘10.27ⴖ
⫺3⬚28⬘15.11ⴖ
⫺3⬚28⬘19.96ⴖ
⫺3⬚28⬘24.80ⴖ
339⬚54⬘05.5ⴖ
341⬚09⬘06.5ⴖ
342⬚24⬘07.5ⴖ
343⬚39⬘08.5ⴖ
344⬚54⬘09.5ⴖ
346⬚09⬘10.5ⴖ
153⬚26⬘51.8ⴖ
154⬚59⬘42.4ⴖ
156⬚33⬘39.0ⴖ
158⬚08⬘39.2ⴖ
159⬚44⬘40.2ⴖ
161⬚21⬘38.9⬘
The Greenwich hour angles for the day were 182⬚34⬘06.00ⴖ at 0h UT,
and 182⬚38⬘53.30ⴖ at 24h UT. The declinations were ⫺3⬚12⬘00.80ⴖ at
0h and ⫺3⬚35⬘16.30ⴖ at 24h. The observer’s latitude and longitude were
scaled from a map as 43⬚15⬘22ⴖ and 90⬚13⬘18ⴖ, respectively, with an
estimated standard error of 1ⴖ for both values. Stopwatch times were
assumed to be correct to within a error of 0.5s. A Roelof’s prism
was used to take pointings on the center of the sun. The target was
535 ft from the observer’s station. The observer’s estimated reading
and pointing errors were 1.2ⴖ and 1.8ⴖ, respectively. The instrument was leveled to within 0.3 div on a level bubble with a sensitivity
of 20ⴖ /div. The target was centered to within an estimated error of
0.003 ft of the station. What is:
(a) the average azimuth of the line and its standard deviation?
(b) the estimated error of the line at 95% level of confidence?
(c) the largest error contributor in the observation?
7.18 The following observations were made on the sun.
126
ERROR PROPAGATION IN ANGLE AND DISTANCE OBSERVATIONS
Pointing
UT Time
Horizontal
Angle
Zenith Angle
1
2
3
4
5
6
13:01:27
13:03:45
13:08:58
13:11:03
13:16:53
13:18:23
179⬚16⬘35ⴖ
179⬚40⬘25ⴖ
180⬚35⬘19ⴖ
180⬚57⬘28ⴖ
182⬚00⬘03ⴖ
182⬚16⬘23ⴖ
56⬚00⬘01ⴖ
55⬚34⬘11ⴖ
54⬚35⬘36ⴖ
54⬚12⬘12ⴖ
53⬚06⬘47ⴖ
52⬚50⬘05ⴖ
The Greenwich hour angles for the day were 178⬚22⬘55.20ⴖ at 0h and
178⬚22⬘58.70ⴖ at 24h. The declinations were 19⬚25⬘44.40⬚ at 0h and
19⬚12⬘18.80ⴖ at 24h. The observer’s latitude and longitude were scaled
from a map as 41⬚18⬘06ⴖ and 75⬚00⬘01ⴖ, respectively, with an estimated
error of 1ⴖ for both. Stopwatch times were assumed to be correct to
within an estimated error of 0.2s. A Roelof’s prism was used to take
pointings on the center of the sun. The target was 335 ft from the
observer’s station. The observer’s estimated reading error was 1.1ⴖ
and the estimated pointing error was 1.6ⴖ. The instrument was leveled to within 0.3 div on a level bubble with a sensitivity of 30ⴖ /div.
The target was centered to within an estimated error of 0.003 ft of
the station. What is:
(a) the average azimuth of the line and its standard deviation?
(b) the estimated error of the line at 95% level of confidence?
(c) the largest error contributor in the observation?
Programming Problems
7.19 Create a computational package that will compute the errors in angle
observations. Use the package to compute the estimated errors for the
angles in Problem 7.11.
7.20 Create a computational package that will compute the errors in EDM
observed distances. Use the package to solve Problem 7.16.
7.21 Create a computational package that will compute the reduced azimuths and their estimated errors from astronomical observations. Use
the package to solve Problem 7.19.
CHAPTER 8
ERROR PROPAGATION IN
TRAVERSE SURVEYS
8.1
INTRODUCTION
Even though the specifications for a project may allow lower accuracies, the
presence of blunders in observations is never acceptable. Thus, an important
question for every surveyor is: How can I tell when blunders are present in
the data? In this chapter we begin to address that question, and in particular,
stress traverse analysis. The topic is discussed further in Chapter 20.
In Chapter 6 it was shown that the estimated error in a function of observations depends on the individual errors in the observations. Generally, observations in horizontal surveys (e.g., traverses) are independent. That is, the
measurement of a distance observation is independent of the azimuth observation. But the latitude and departure of a line, which are computed from the
distance and azimuth observations, are not independent. Figure 8.1 shows the
effects of errors in distance and azimuth observations on the computed latitude
and departure. In the figure it can be seen that there is correlation between
the latitude and departure; that is, if either distance or azimuth observation
changes, it causes changes in both latitude and departure.
Because the observations from which latitudes and departures are computed are assumed to be independent with no correlation, the SLOPOV approach [Equation (6.16)] can be used to determine the estimated error in these
computed values. However, for proper computation of estimated errors in
functions that use these computed values (i.e., latitudes and departures), the
effects of correlation must be considered, and thus the GLOPOV approach
[Equation (6.13)] will be used.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
127
128
ERROR PROPAGATION IN TRAVERSE SURVEYS
Figure 8.1 Latitude and departure uncertainties due to (a) the distance error (D) and
(b) the azimuth error Az. Note that if either the distance or azimuth changes, both the
latitude and departure of the course are affected.
8.2 DERIVATION OF ESTIMATED ERROR IN LATITUDE
AND DEPARTURE
When computing the latitude and departure of a line, the following wellknown equations are used:
Lat ⫽ D cos Az
(8.1)
Dep ⫽ D sin Az
where Lat is the latitude, Dep the departure, Az the azimuth, and D the
horizontal length of the line. To derive the estimated error in the line’s latitude
or departure, the following partial derivatives from Equation (8.1) are required
in using Equation (6.16):
⭸Lat
⫽ cos Az
⭸D
⭸Dep
⫽ sin Az
⭸D
⭸Lat
⫽ ⫺D sin Az
⭸Az
⭸Dep
⫽ D cos Az
⭸Az
(8.2)
Example 8.1 A traverse course has a length of 456.87 0.02 ft and an
azimuth of 23⬚35⬘26ⴖ 9ⴖ. What are the latitude and departure and their
estimated errors?
SOLUTION Using Equation (8.1), the latitude and departure of the course
are
Lat ⫽ 456.87 cos(23⬚35⬘26ⴖ) ⫽ 418.69 ft
Dep ⫽ 456.87 sin(23⬚35⬘26ⴖ) ⫽ 182.84 ft
The estimated errors in these values are solved using matrix Equation (6.16)
as
8.3
⌺Lat,Dep ⫽
DERIVATION OF ESTIMATED STANDARD ERRORS IN COURSE AZIMUTHS
冤
⭸Lat
⭸D
⭸Dep
⭸D
冥冋
⭸Lat
⭸Az
⭸Dep
⭸Az
2D
0
册冤
2
Az
0
⭸Lat
⭸D
⭸Lat
⭸Az
冥
⭸Dep
⭸D
⭸Dep
⭸Az
⫽
冋
129
册
2Lat
Lat,Dep
2Dep
Lat,Dep
Substituting partial derivatives into the above yields
⌺Lat,Dep ⫽
冋
册冋
cos Az ⫺D sin Az
sin Az
D cos Az
0.022
0
0
(9ⴖ / )2
册冋
cos Az
⫺D sin Az
册
sin Az
D cos Az
(8.3)
Entering in the appropriate numerical values into Equation (8.3), the covariance matrix is
⌺Lat,Dep ⫽
冋
冋
册冋
0.9167 ⫺456.87(0.4002)
0.4002
456.87(0.9164)
0.9167
⫺456.87(0.4002)
0.0004
0
册
册
0
(9ⴖ / )2
0.4002
456.87(0.9164)
from which
⌺Lat,Dep ⫽
冋
册
0.00039958 0.00000096
0.00000096 0.00039781
(8.4)
In Equation (8.4), 211 is the variance of the latitude, 222 the variance of
the departure, and 12 and 21 their covariances. Thus, the standard errors are
Lat ⫽ 兹211 ⫽ 兹0.00039958 ⫽ 0.020 ft
and
Dep ⫽ 兹222 ⫽ 兹0.00039781 ⫽ 0.020 ft
Note that the off-diagonal of ⌺Lat,Dep is not equal to zero, and thus the computed values are correlated as illustrated in Figure 8.1.
8.3 DERIVATION OF ESTIMATED STANDARD ERRORS IN
COURSE AZIMUTHS
Equation (8.1) is based on the azimuth of a course. In practice, however,
traverse azimuths are normally computed from observed angles rather than
being measured directly. Thus, another level of error propagation exists in
calculating the azimuths from angular values. In the following analysis, con-
130
ERROR PROPAGATION IN TRAVERSE SURVEYS
sider that angles to the right are observed and that azimuths are computed in
a counterclockwise direction successively around the traverse using the
formula
AzC ⫽ AzP ⫹ 180⬚ ⫹ i
(8.5)
where AzC is the azimuth for the current course, AzP the previous course
azimuth, and i the appropriate interior angle to use in computing the current
course azimuth. By applying Equation (6.18), the error in the current azimuth,
AzC, is
2
AzC ⫽ 兹Az
⫹ 2i
P
(8.6)
In Equation (8.6) is the error in the appropriate interior angle used in
computation of the current azimuth, and the other terms are as defined previously. This equation is also valid for azimuth computations going clockwise
around the traverse. The proof of this is left as an exercise.
8.4 COMPUTING AND ANALYZING POLYGON TRAVERSE
MISCLOSURE ERRORS
From elementary surveying it is known that the following geometric constraints exist for any closed polygon-type traverse:
兺 interior ∠’s ⫽ (n ⫺ 2) ⫻ 180⬚
(8.7)
兺 Lat ⫽ 兺 Dep ⫽ 0
(8.8)
Deviations from these conditions, normally called misclosures, can be
calculated from the observations of any traverse. Statistical analyses can then
be performed to determine the acceptability of the misclosures and check for
the presence of blunders in the observations. If blunders appear to be present,
the measurements must be rejected and the observations repeated. The following example illustrates methods of making these computations for any
closed polygon traverse.
Example 8.2 Compute the angular and linear misclosures for the traverse
illustrated in Figure 8.2. The observations for the traverse are given in Table
8.1. Determine the estimated misclosure errors at the 95% confidence level,
and comment on whether or not the observations contain blunders.
SOLUTION
Angular check: First the angular misclosure is checked to see if it is within
the tolerances specified. From Equation (6.18), and using the standard devi-
8.4
COMPUTING AND ANALYZING POLYGON TRAVERSE MISCLOSURE ERRORS
131
Figure 8.2 Closed polygon traverse.
ations given in Table 8.1, the angular sum should have an error within
2
2
兹2∠1 ⫹ ∠2
⫹ ⫹ ∠n
68.3% of the time. Since the angles were measured
four times, each computed mean has three degrees of freedom, and the appropriate t value from Table D.3 (the t distribution) is t0.025,3, which equals
3.183. This is a two-tailed test since we are looking for the range that is
statistically equal to zero at the level of confidence selected. If this range
contains the actual misclosure, there is no statistical reason to believe that the
observations contain a blunder. In this case, the angular misclosure at a 95%
confidence level is estimated as
兺
Angles
⫽ 3.183兹3.52 ⫹ 3.12 ⫹ 3.62 ⫹ 3.12 ⫹ 3.92 ⫽ 24.6ⴖ
Using the summation of the angles in Table 8.1, the actual angular misclosure
in this problem is
540⬚00⬘19ⴖ ⫺ (5 ⫺ 2)180⬚ ⫽ 19ⴖ
Thus, the actual angular misclosure for the traverse (19ⴖ) is within its estimated range of error and there is no reason to believe that a blunder exists
in the angles.
TABLE 8.1 Distance and Angle Observations for Figure 8.2
Station
Sighted
Distance (ft)
S (ft)
BS
Occupied
FS
Angle a
S
A
B
C
D
E
B
C
D
E
A
1435.67
856.94
1125.66
1054.54
756.35
0.020
0.020
0.020
0.020
0.020
E
A
B
C
D
A
B
C
D
E
B
C
D
E
A
110⬚24⬘40ⴖ
87⬚36⬘14ⴖ
125⬚47⬘27ⴖ
99⬚57⬘02ⴖ
116⬚14⬘56ⴖ
540⬚00⬘19ⴖ
3.5ⴖ
3.1ⴖ
3.6ⴖ
3.1ⴖ
3.9ⴖ
a
Each angle was measured with four repetitions.
132
ERROR PROPAGATION IN TRAVERSE SURVEYS
Azimuth computation: In this problem, no azimuth is given for the first
course. To solve the problem, however, the azimuth of the first course can be
assumed as 0⬚00⬘00ⴖ and to be free of error. This can be done even when the
initial course azimuth is observed, since only geometric closure on the traverse is being checked, not the orientation of the traverse. For the data of Table
8.1, and using Equations (8.5) and (8.6), the values for the course azimuths
and their estimated errors are computed and listed in Table 8.2.
Computation of estimated linear misclosure: Equation (6.13) properly accounts for correlation in the latitude and departure when computing the linear
misclosure of the traverse. Applying the partial derivatives of Equation (8 .2)
to the latitudes and departures, the Jacobian matrix, A, has the form
冤
cos AzAB ⫺AB sin AzAB
0
0
sin AzAB AB cos AzAB
0
0
0
0
cos AzBC ⫺BC sin AzBC
A⫽
0
0
sin AzBC BC cos AzBC
⯗
⯗
0
0
0
0
0
0
0
0
0
0
0
0
冥
0
0
0
0
cos AzEA ⫺EA sin AzEA
sin AzEA EA cos AzEA
(8.9)
Because the lengths and angles were measured independently, they are uncorrelated. Thus, the appropriate covariance matrix, ⌺, for solving this problem using Equation (6.16) is
2AB
0
0
0
冉 冊
AzAB
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
冉 冊
0
0
0
0
0
0
0
0
0
0
0
0
0
2CD
冉 冊
0
0
0
0
0
0
0
0
0
2DE
冉 冊
0
0
0
0
0
2EA
2
2BC
0
0
0
0
0
0
0
AzBC
2
⌺⫽
0
0
0
0
0
0
0
0
0
0
0
AzCD
2
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
AzDE
0
2
0
冉 冊
AzEA
2
(8.10)
8.4
133
COMPUTING AND ANALYZING POLYGON TRAVERSE MISCLOSURE ERRORS
TABLE 8.2 Estimated Errors in the Computed Azimuths of Figure 8.2
From
To
Azimuth
Estimated Error
A
B
C
D
E
B
C
D
E
A
0⬚00⬘00ⴖ
267⬚36⬘14ⴖ
213⬚23⬘41ⴖ
133⬚20⬘43ⴖ
69⬚35⬘39ⴖ
0ⴖ
3.1ⴖ
兹3.12 ⫹ 3.62 ⫽ 4.8ⴖ
兹4.82 ⫹ 3.22 ⫽ 5.7ⴖ
兹5.72 ⫹ 3.92 ⫽ 6.9ⴖ
Substituting numerical values for this problem into Equations (8.9) and (8.10),
the covariance matrix, ⌺Lat,Dep, is computed for the latitudes and departures
A⌺AT, or
⌺Lat,Dep ⫽
0.00040
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0.00017
0.00002
0
0
0
0
0
0
0
0
0.00002
0.00040
0
0
0
0
0
0
0
0
0
0
0.00049
0.00050
0
0
0
0
0
0
0
0
0
0
0
0
0.00050
0
0.00060
0
0
0.00064
0
⫺0.00062
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
⫺0.00062
0
0.00061
0
0
0.00061
0
0.00034
0
0
0
0
0
0
0
0
0.00034
0.00040
(8.11)
By taking the square roots of the diagonal elements in the ⌺Lat,Dep matrix
[Equation (8.11)], the errors for the latitude and departure of each course are
found. That is, the estimated error in the latitude for course BC is the square
root of the (3,3) element in Equation (8.11), and the estimated error in the
departure of BC is the square root of the (4,4) element. In a similar fashion,
the estimated errors in latitude and departure can be computed for any other
course.
The formula for determining the linear misclosure of a closed polygon
traverse is
LC ⫽ 兹(LatAB ⫹ LatBC ⫹ ⫹ LatEA)2 ⫹ (DepAB ⫹ DepBC ⫹ ⫹ DepEA)2
(8.12)
where LC is the linear misclosure. To determine the estimated error in the
linear misclosure, Equation (6.16) is applied to the linear misclosure formula
(8.12). The necessary partial derivatives from Equation (8.12) for substitution
into Equation (6.16) must first be determined. The partial derivatives with
respect to the latitude and departure of course AB are
134
ERROR PROPAGATION IN TRAVERSE SURVEYS
⭸LC
⫽
⭸LatAB
兺 Lats
LC
⭸LC
⫽
⭸DepAB
兺 Deps
(8.13)
LC
Notice that these partial derivatives are independent of the course. Also,
the other courses have the same partial derivatives as given by Equation
(8.13), and thus the Jacobian matrix for Equation (6.16) has the form
A⫽
冋
兺 Lats 兺 Deps 兺 Lats 兺 Deps
LC
LC
LC
LC
册
兺 Lats 兺 Deps
LC
LC
(8.14)
As shown in Table 8.3, the sum of the latitudes is ⫺0.083, the sum of the
departures is 0.022, and LC ⫽ 0.086 ft. Substituting these values into Equation
(8.14), which in turn is substituted into Equation (6.16), yields
⌺LC ⫽ [⫺0.9674
0.2531 ⫺0.9674
0.2531 ⫺0.9674
冤 冥
0.2531]
⫺0.9674
0.2531
⫺0.9674
⫻ ⌺Lat,Dep
0.2531 ⫽ [0.00226]
⯗
⫺0.9674
0.2531
(8.15)
In Equation (8.15), ⌺LC is a single-element covariance matrix that is the
variance of the linear closure and can be called 2LC. Also, ⌺Lat,Dep is the matrix
given by Equation (8.11). To compute the E95 confidence interval, a t value
from Table D.3 (the t distribution) must be used with ␣ ⫽ 0.025 and 3 degrees
TABLE 8.3 Latitudes and Departures for
Example 8.2
Course
Latitude
AB
BC
CD
DE
EA
1435.670
⫺35.827
⫺939.812
⫺723.829
263.715
⫺0.083
Departure
0
⫺856.191
⫺619.567
766.894
708.886
0.022
LC ⫽ 兹(⫺0.083)2 ⫹ (0.022)2 ⫽ 0.086 ft
8.5
COMPUTING AND ANALYZING LINK TRAVERSE MISCLOSURE ERRORS
135
of freedom.1 The misclosure estimated for a traverse at a 1 ⫺ ␣ level of
confidence is t␣ / 2,3LC. Again, we are checking to see if the traverse misclosure falls within a range of errors that are statistically equal to zero. This
requires placing ␣ /2 into the upper and lower tails of the distribution. Thus,
the error estimated in the traverse closure at a 95% level of confidence is
LC ⫽ t0.025,3兹21 ⫽ 3.183兹0.00226 ⫽ 0.15 ft
This value is well above the actual traverse linear misclosure of 0.086 ft, and
thus there is no reason to believe that the traverse contains any blunders.
In Example 8.2 we failed to reject the null hypothesis; that is, there was
no statistical reason to believe that there were errors in the data. However, it
is important to remember that this does necessarily imply that the observations
are error-free. There is always the possibility of a Type II error. For example,
if the computations were supposed to be performed on a map projection grid2
but the observations were not reduced, the traverse would still close within
acceptable tolerances. However, the results computed would be incorrect since
all the distances would be either too long or too short. Another example of
an undetectable systematic error is an incorrectly entered EDM–reflector constant (see Problem 2.17). Again all the distances observed would be either
too long or too short, but the traverse misclosure would still be within acceptable tolerances.
Surveyors must always be aware of instrumental systematic errors and
follow proper field and office procedures to remove these errors. As discussed,
simply passing a statistical test does not imply directly that the observations
are error- or mistake-free. However, when the test fails, only a Type I error
can occur at an ␣ level of confidence. Depending on the value of ␣, a failed
test can be a strong indicator of problems within the data.
8.5 COMPUTING AND ANALYZING LINK TRAVERSE
MISCLOSURE ERRORS
As illustrated in Figure 8.3, a link traverse begins at one station and ends on
a different one. Normally, they are used to establish the positions of inter1
A closed polygon traverse has 2(n ⫺ 1) unknown coordinates with 2n ⫹ 1 observations, where
n is the number of traverse sides. Thus, the number of degrees of freedom in a simple closed
traverse is always 2n ⫹ 1 ⫺ 2(n ⫺ 1) ⫽ 3. For a five-sided traverse there are five angle and five
distance observations plus one azimuth. Also, there are four stations each having two unknown
coordinates, thus 11 ⫺ 8 ⫽ 3 degrees of freedom.
2
Readers who wish to familiarize themselves with map projection computations should refer to
Appendix F and the CD that accompanies this book.
136
ERROR PROPAGATION IN TRAVERSE SURVEYS
Figure 8.3 Closed link traverse.
mediate stations, as in A through D of the figure. The coordinates at the
endpoints, stations 1 and 2 of the figure, are known. Angular and linear misclosures are also computed for these types of traverse, and the resulting values
are used as the basis for accepting or rejecting the observations. Example 8.3
illustrates the computational methods.
Example 8.3 Compute the angular and linear misclosures for the traverse
illustrated in Figure 8.3. The data observed for the traverse are given in Table
8.4. Determine the estimated misclosures at the 95% confidence level, and
comment on whether or not the observations contain blunders.
SOLUTION
Angular misclosure: In a link traverse, angular misclosure is found by
computing initial azimuths for each course and then subtracting the final com-
TABLE 8.4 Data for Link Traverse in Example 8.3
Distance observations
Control stations
From
To
Distance (ft)
S (ft)
Station
X (ft)
Y (ft)
1
A
B
C
D
A
B
C
D
2
1069.16
933.26
819.98
1223.33
1273.22
0.021
0.020
0.020
0.021
0.021
1
2
1248.00
4873.00
3979.00
3677.00
Azimuth observations
Angle observations
BS
1
A
B
C
Occ
FS
Angle
S (ⴖ)
A
B
C
D
B
C
D
2
66⬚16⬘35ⴖ
205⬚16⬘46ⴖ
123⬚40⬘19ⴖ
212⬚00⬘55ⴖ
4.9
5.5
5.1
4.6
From
To
1
D
A
2
Azimuth
S (ⴖ)
197⬚04⬘47ⴖ 4.3
264⬚19⬘13ⴖ 4.1
8.5
COMPUTING AND ANALYZING LINK TRAVERSE MISCLOSURE ERRORS
137
TABLE 8.5 Computed Azimuths and Their
Uncertainties
Course
Azimuth
( ⴖ)
1A
AB
BC
CD
D2
197⬚04⬘47ⴖ
83⬚21⬘22ⴖ
108⬚38⬘08ⴖ
52⬚18⬘27ⴖ
84⬚19⬘22ⴖ
4.3
6.5
8.5
9.9
11.0
puted azimuth from its given counterpart. The initial azimuths and their estimated errors are computed using Equations (8.5) and (8.6) and are shown
in Table 8.5.
The difference between the azimuth computed for course D2 (84⬚19⬘22ⴖ)
and its actual value (264⬚19⬘13ⴖ ⫺ 180⬚) is ⫹9ⴖ. Using Equation (6.18), the
estimated error in the difference is 兹11.02 ⫹ 4.12 ⫽ 11.7ⴖ, and thus there
is no reason to assume that the angles contain blunders.
Linear misclosure: First the actual traverse misclosure is computed using
Equation (8.1). From Table 8.6, the total change in latitude for the traverse
is ⫺302.128 ft and the total change in departure is 3624.968 ft. From the
control coordinates, the cumulative change in X and Y coordinate values is
⌬X ⫽ X2 ⫺ X1 ⫽ 4873.00 ⫺ 1248.00 ⫽ 3625.00 ft
⌬Y ⫽ Y2 ⫺ Y1 ⫽ 3677.00 ⫺ 3979.00 ⫽ ⫺302.00 ft
The actual misclosures in departure and latitude are computed as
⌬Dep ⫽
⌬Lat ⫽
兺 Dep ⫺ (X2 ⫺ X1) ⫽ 3624.968 ⫺ 3625.00 ⫽ ⫺0.032
兺 Lat ⫺ (Y2 ⫺ Y1) ⫽ ⫺302.128 ⫺ (⫺302.00) ⫽ ⫺0.128
TABLE 8.6 Computed Latitudes and Departures
Course
Latitude (ft)
Departure (ft)
1A
AB
BC
CD
D2
⫺1022.007
⫺314.014
107.976
⫺262.022
747.973
125.952
⫺302.128
926.993
776.989
968.025
1266.975
3624.968
(8.16)
138
ERROR PROPAGATION IN TRAVERSE SURVEYS
In Equation (8.16), ⌬Dep represents the misclosure in departure and ⌬Lat
represents the misclosure in latitude. Thus, the linear misclosure for the traverse is
LC ⫽ 兹⌬Lat2 ⫹ ⌬Dep2 ⫽ 兹(⫺0.128)2 ⫹ (⫺0.032)2 ⫽ 0.132 ft
(8.17)
Estimated misclosure for the traverse: Following procedures similar to
those described earlier for polygon traverses, the estimated misclosure in this
link traverse is computed. The Jacobian matrix of the partial derivative for
the latitude and departure with respect to distance and angle observations is
冤
cos Az1A ⫺1A sin AzA
0
0
sin Az1A 1A cos Az1A
0
0
0
0
cos AzAB ⫺AB sin AzAB
A⫽
0
0
sin AzAB AB cos AzAB
⯗
⯗
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
cos AzD2 ⫺D2 sin AzD2
sin AzD2 D2 cos AzD2
冥
(8.18)
Similarly, the corresponding covariance matrix in Equation (6.16) has
the form
21A
0
0
⫽
0
冉 冊
Az1A
0
0
0
0
0
0
0
0
0
2AB
0
0
0
0
0
0
0
2D2
0
2
冉 冊
AzAB
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
2
0
(8.19)
冉 冊
AzD2
2
Substituting the appropriate numerical values into Equations (8.18) and (8.19)
and applying Equation (6.16), the covariance matrix is
8.5
COMPUTING AND ANALYZING LINK TRAVERSE MISCLOSURE ERRORS
139
Lat,Dep ⫽
0.00045 ⫺0.00016
0
0
0
0
0
0
0
0
0.00049
0
0
0
0
0
0
0
0
0
0
0.00086 ⫺0.00054
0
0
0
0
0
0
0
0
⫺0.00054 0.00041
0
0
0
0
0
0
0
0
0
0
0.00107 0.00023
0
0
0
0
0
0
0
0
0.00023 0.00048
0
0
0
0
0
0
0
0
0
0
0.00234 ⫺0.00147
0
0
0
0
0
0
0
0
⫺0.00147 0.00157
0
0
0
0
0
0
0
0
0
0
0.00453 ⫺0.00041
0
0
0
0
0
0
0
0
⫺0.00041 0.00048
⫺0.00016
To estimate the error in the traverse misclosure, Equation (6.16) must be
applied to Equation (8.18). As was the case for closed polygon traverse, the
terms of the Jacobian matrix are independent of the course for which they
are determined, and thus the Jacobian matrix has the form
A⫽
冋
Lat
LC
Dep
LC
Lat
LC
Dep
LC
Lat
LC
册
Dep
LC
(8.20)
Following procedures similar to those used in Example 8.2, the estimated
standard error in the misclosure of the link traverse is
LC ⫽ ALat,Dep AT ⫽ [0.00411]
From these results and using a t value from Table D.3 for 3 degrees of
freedom, the estimated linear misclosure error for a 95% level of confidence
is
95% ⫽ 3.183兹0.00411 ⫽ 0.20 ft
Because the actual misclosure of 0.13 ft is within the range of values that are
statistically equal to zero at the 95% level (0.20 ft), there is no reason to
believe that the traverse observations contain any blunders. Again, this test
does not remove the possibility of a Type II error occurring.
This example leads to an interesting discussion. When using traditional
methods of adjusting link traverse data, such as the compass rule, the control
is assumed to be perfect. However, since control coordinates are themselves
derived from observations, they contain errors that are not accounted for in
these computations. This fact is apparent in Equation (8.20), where the coordinate values are assumed to have no error and thus are not represented.
140
ERROR PROPAGATION IN TRAVERSE SURVEYS
These equations can easily be modified to consider the control errors, but this
is left as an exercise for the student.
One of the principal advantages of the least squares adjustment method is
that it allows application of varying weights to the observations, and control
can be included in the adjustment with appropriate weights. A full discussion
of this subject is presented in Section 21.6.
8.6
CONCLUSIONS
In this chapter, propagation of observational errors through traverse computations has been discussed. Error propagation is a powerful tool for the surveyor, enabling an answer to be obtained for the question: What is an
acceptable traverse misclosure? This is an example of surveying engineering.
Surveyors are constantly designing measurement systems and checking their
results against personal or legal standards. The subjects of error propagation
and detection of measurement blunders are discussed further in later chapters.
PROBLEMS
8.1
Show that Equation (8.6) is valid for clockwise computations about a
traverse.
8.2
Explain the significance of the standard error in the azimuth of the first
course of a polygon traverse.
8.3
Given a course with an azimuth of 105⬚27⬘44ⴖ with an estimated error
of 5ⴖ and a distance of 638.37 ft with an estimated error of 0.02
ft, what are the latitude and departure and their estimated errors?
8.4
Given a course with an azimuth of 272⬚14⬘08ⴖ with an estimated error
of 9.2ⴖ and a distance of 215.69 ft with an estimated error of 0.016
ft, what are the latitude and departure and their estimated errors?
8.5
Given a course with an azimuth of 328⬚49⬘06ⴖ with an estimated error
of 4.4ⴖ and a distance of 365.977 m with an estimated error of 6.5
mm, what are the latitude and departure and their estimated errors?
8.6
Given a course with an azimuth of 44⬚56⬘22ⴖ with an estimated error
of 6.7ⴖ and a distance of 138.042 m with an estimated error of 5.2
mm, what are the latitude and departure and their estimated errors?
8.7
A traverse meets statistical closures at the 95% level of confidence. In
your own words, explain why this does not necessarily imply that the
traverse is without error.
8.8
A polygon traverse has the following angle measurements and related
standard deviations. Each angle was observed twice (one direct and
PROBLEMS
141
one reverse). Do the angles meet acceptable closure limits at a 95%
level of confidence?
8.9
Backsight
Occupied
Foresight
Angle
S
A
B
C
D
B
C
D
A
C
D
A
B
107⬚53⬘31ⴖ
81⬚56⬘44ⴖ
92⬚34⬘28ⴖ
77⬚35⬘39ⴖ
2.2ⴖ
2.4ⴖ
3.2ⴖ
2.8ⴖ
Given an initial azimuth for course AB of 36⬚34⬘25ⴖ, what are the
azimuths and their estimated standard errors for the remaining three
courses of Problem 8.8?
8.10 Using the distances listed in the following table and the data from
Problems 8.8 and 8.9, compute:
(a) the misclosure of the traverse.
(b) the estimated misclosure error.
(c) the 95% misclosure error.
From
To
Distance (ft)
S (ft)
A
B
C
D
B
C
D
A
211.73
302.49
254.48
258.58
0.016
0.017
0.016
0.016
8.11 Given the traverse misclosures in Problem 8.10, does the traverse meet
acceptable closure limits at a 95% level of confidence? Justify your
answer statistically.
8.12 Using the data for the link traverse listed below, compute:
(a) the angular misclosure and its estimated error.
(b) the misclosure of the traverse.
(c) the estimated misclosure error.
(d) the 95% error in the traverse misclosure.
Angle observations
Distance observations
From To Distance (m)
W
X
Y
X
Y
Z
185.608
106.821
250.981
(m)
0.0032
0.0035
0.0028
Back Occ For
W
X
X
Y
Y
Z
Angle
(ⴖ)
86⬚27⬘45ⴖ 2.5
199⬚29⬘46ⴖ 3.2
142
ERROR PROPAGATION IN TRAVERSE SURVEYS
Control azimuths
Control stations
From
To
Azimuth
(ⴖ)
Station
Easting (m)
Northing (m)
W
Y
X
Z
132⬚26⬘15ⴖ
58⬚23⬘56ⴖ
9ⴖ
8ⴖ
W
Z
10,000.000
10,417.798
5000.000
5089.427
8.13 Does the link traverse of Problem 8.12 have acceptable traverse misclosure at a 95% level of confidence? Justify your answer statistically.
8.14 Following are the length and azimuth data for a city lot survey.
Course
Distance (ft)
D (ft)
Azimuth
Az
AB
BC
CD
DA
134.58
156.14
134.54
156.10
0.02
0.02
0.02
0.02
83⬚59⬘54ⴖ
353⬚59⬘44ⴖ
263⬚59⬘54ⴖ
174⬚00⬘04ⴖ
0ⴖ
20ⴖ
28ⴖ
35ⴖ
Compute:
(a) the misclosure of the traverse.
(b) the estimated misclosure error.
(c) the 95% misclosure error.
(d) Does the traverse meet acceptable 95% closure limits? Justify your
response with statistically.
8.15 Repeat Problem 8.14 using the data from Problems 7.11 and 7.15.
8.16 Repeat Problem 8.14 using the data from Problems 7.12 and 7.15.
8.17 A survey produces the following set of data. The angles were obtained
from the average of four measurements (two direct and two reverse)
made with a total station. The estimated uncertainties in the observations are
DIN ⫽ 3ⴖ
t ⫽ 0.010 ft
i ⫽ 0.003 ft
The EDM instrument has a specified accuracy of (3 mm ⫹ 3 ppm).
Distance observations
Angle observations
From
To
Distance (ft)
Back
Occ.
For
Angle
1
2
3
4
5
2
3
4
5
1
999.99
801.55
1680.03
1264.92
1878.82
5
1
2
3
4
1
2
3
4
5
2
3
4
5
1
191⬚40⬘12ⴖ
56⬚42⬘22ⴖ
122⬚57⬘10ⴖ
125⬚02⬘11ⴖ
43⬚38⬘10ⴖ
PROBLEMS
Control azimuths
143
Control stations
From
To
Azimuth
( ⴖ)
Station
Easting (ft)
Northing (ft)
1
2
216⬚52⬘11ⴖ
3
1
1000.00
1000.00
Compute:
(a) the estimated errors in angles and distances.
(b) the angular misclosure and its 95% probable error.
(c) the misclosure of the traverse.
(d) the estimated misclosure error and its 95% value.
(e) Did the traverse meet acceptable closures? Justify your response
statistically.
8.18 Develop new matrices for the link traverse of Example 8.3 that considers the errors in the control. Assume control coordinate standard
errors of x ⫽ y ⫽ 0.05 ft for both stations 1 and 2, and use these
new matrices to compute:
(a) the estimated misclosure error.
(b) the 95% misclosure error.
(c) Compare these results with those in the examples.
Programming Problems
8.19 Develop a computational package that will compute the course azimuths and their estimated errors given an initial azimuth and measured
angles. Use this package to answer Problem 8.9.
8.20 Develop a computational package that will compute estimated traverse
misclosure error given course azimuths and distances and their estimated errors. Use this package to answer Problem 8.10.
8.21 Develop a computational package that will compute estimated traverse
misclosure errors given the data of Problem 8.17.
CHAPTER 9
ERROR PROPAGATION IN
ELEVATION DETERMINATION
9.1
INTRODUCTION
Differential and trigonometric leveling are the two most commonly employed
procedures for finding elevation differences between stations. Both of these
methods are subject to systematic and random errors. The primary systematic
errors include Earth curvature, atmospheric refraction, and instrument maladjustment. The effects of these systematic errors can be minimized by following proper field procedures. They can also be modeled and corrected for
computationally. The random errors in differential and trigonometric leveling
occur in instrument leveling, distance observations, and reading graduated
scales. These must be treated according to the theory of random errors.
9.2
SYSTEMATIC ERRORS IN DIFFERENTIAL LEVELING
During the differential leveling process, sight distances are held short, and
equal, to minimize the effects of systematic errors. Still, it should always be
assumed that these errors are present in differential leveling observations, and
thus corrective field procedures should be followed to minimize their effects.
These procedures are the subjects of discussions that follow.
9.2.1
Collimation Error
Collimation error occurs when the line of sight of an instrument is not truly
horizontal and is minimized by keeping sight distances short and balanced.
144
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
9.2
SYSTEMATIC ERRORS IN DIFFERENTIAL LEVELING
145
Figure 9.1 Collimation error in differential leveling.
Figure 9.1 shows the effects of a collimation error. For an individual setup,
the resulting error in an elevation difference due to collimation is
(9.1)
eC ⫽ D1␣ ⫺ D2␣
where eC is the error in elevation due to the presence of a collimation error,
D1 and D2 the distances to the backsight and foresight rods, respectively, and
␣ the amount of collimation error present at the time of the observation expressed in radian units. Applying Equation (9.1), the collimation error for a
line of levels can be expressed as
eC ⫽ ␣[(D1 ⫺ D2) ⫹ (D3 ⫺ D4) ⫹ 䡠 䡠 䡠 ⫹ (Dn⫺1 ⫺ Dn)]
(9.2)
where D1, D3, . . . , Dn⫺1 are the backsight distances and D2, D4, . . . , Dn are
the foresight distances. If the backsight and foresight distances are grouped
in Equation (9.2),
eC ⫽ ␣
冉冘
DBS ⫺
冘D
冊
FS
(9.3)
The collimation error determined from Equation (9.3) is treated as a correction
and thus subtracted from the observed elevation difference to obtain the corrected value.
Example 9.1 A level having a collimation error of 0.04 mm/m is used on
a level line where the backsight distances sum to 863 m and the foresight
distances sum to 932 m. If the elevation difference observed for the line is
22.865 m, what is the corrected elevation difference?
146
ERROR PROPAGATION IN ELEVATION DETERMINATION
SOLUTION Using Equation (9.3), the error due to collimation is
eC ⫽ 0.00004(863 ⫺ 932) m ⫽ ⫺0.0028 m
Thus, the corrected elevation difference is
22.865 ⫺ (⫺0.00276) ⫽ 22.868 m
9.2.2
Earth Curvature and Refraction
As the line of sight extends from an instrument, the level surface curves down
and away from it. This condition always causes rod readings to be too high.
Similarly, as the line of sight extends from the instrument, refraction bends
it toward the Earth surface, causing readings to be too low. The combined
effect of Earth curvature and refraction on an individual sight always causes
a rod reading to be too high by an amount
冉 冊
2
D
1000
hCR ⫽ CR
(9.4)
where hCR is the error in the rod reading (in feet or meters), CR is 0.0675
when D is in units of meters or 0.0206 when D is in units of feet, and D is
the individual sight distance.
The effect of this error on a single elevation difference is minimized by
keeping backsight and foresight distances short and equal. For unequal sight
distances, the resulting error is expressed as
冉 冊
eCR ⫽ CR
D1
1000
2
冉 冊
⫺ CR
D2
1000
2
(9.5)
where eCR is the error due to Earth curvature and refraction on a single elevation difference. Factoring common terms in Equation (9.5) yields
eCR ⫽
CR
(D2 ⫺ D22)
10002 1
(9.6)
To get the corrected value, the curvature and refraction error computed from
Equation (9.6) is treated as a correction and thus subtracted from the elevation
difference observed.
9.2
147
SYSTEMATIC ERRORS IN DIFFERENTIAL LEVELING
Example 9.2 An elevation difference between two stations on a hillside is
determined to be 1.256 m. What would be the error in the elevation difference
and the corrected elevation difference if the backsight distance were 100 m
and the foresight distance only 20 m?
SOLUTION Substituting the distances into Equation (9.6) and using CR ⫽
0.0675 gives us
eCR ⫽
0.0675
(1002 ⫺ 202) ⫽ 0.0006 m
10002
From this, the corrected elevation difference is
⌬h ⫽ 1.256 ⫺ 0.0006 ⫽ 1.255 m
For a line of differential leveling, the combined effect of this error is
eCR ⫽
CR
(D2 ⫺ D22 ⫹ D32 ⫺ D42 ⫹ 䡠 䡠 䡠)
10002 1
(9.7)
Regrouping backsight and foresight distances, Equation (9.7) becomes
eCR ⫽
CR
10002
冉冘
2
DBS
⫺
冘D
冊
2
FS
(9.8)
The error due to refraction caused by the vertical gradient of temperature
can be large when sight lines are allowed to pass through the lower layers of
the atmosphere. Since observing the temperature gradient along the sight line
would be cost prohibitive, a field procedure is generally adopted that requires
all sight lines to be at least 0.5 m above the ground. This requirement eliminates the lower layers of the atmosphere, where refraction is difficult to
model.
9.2.3 Combined Effects of Systematic Errors on
Elevation Differences
With reference to Figure 9.1, and by combining Equations (9.1) and (9.5), a
corrected elevation difference, ⌬h, for one instrument setup is
148
ERROR PROPAGATION IN ELEVATION DETERMINATION
⌬h ⫽ (r1 ⫺ r2) ⫺ (D1␣ ⫺ D2␣) ⫺
CR
(D2 ⫺ D22)
10002 1
(9.9)
where r1 is the backsight rod reading, r2 the foresight rod reading, and the
other terms are as defined previously.
9.3
RANDOM ERRORS IN DIFFERENTIAL LEVELING
Differential leveling is subject to several sources of random errors. Included
are errors in leveling the instrument and in reading the rod. The sizes of these
errors are affected by atmospheric conditions, the quality of the optics of the
telescope, the sensitivity of the level bubble or compensator, and the graduation scale on the rods. These errors are discussed below.
9.3.1
Reading Errors
The estimated error in rod readings is usually expressed as a ratio of the
estimated standard error in the rod reading per unit sight distance length. For
example, if an observer’s ability to read a rod is within Ⳳ0.005 ft per 100
ft, then r / D is Ⳳ0.005/100 ⫽ Ⳳ0.00005 ft/ft. Using this, rod reading errors
for any individual sight distance D can be estimated as
r ⫽ Dr / D
(9.10)
where r / D is the estimated error in the rod reading per unit length of sight
distance and D is the length of the sight distance.
9.3.2
Instrument Leveling Errors
The estimated error in leveling for an automatic compensator or level vial is
generally given in the technical data for each instrument. For precise levels,
this information is usually listed in arc seconds or as an estimated elevation
error for a given distance. As an example, the estimated error may be listed
as Ⳳ1.5 mm/km, which corresponds to Ⳳ1.5/1,000,000 ⫻ ⫽ Ⳳ0.3ⴖ. A
precise level will usually have a compensator accuracy or setting accuracy
between Ⳳ0.1ⴖ and Ⳳ0.2ⴖ, whereas for a less precise level, the value may be
as high as Ⳳ10ⴖ.
9.3.3
Rod Plumbing Error
Although a level rod that is held nonvertical always causes the reading to be
too high, this error will appear random in a leveling network, due to its
presence in all backsight and foresight distances of the network. Thus, the
rod plumbing error should be modeled when computing the standard error in
9.3
RANDOM ERRORS IN DIFFERENTIAL LEVELING
149
Figure 9.2 Novertical level rod.
an elevation difference. With reference to Figure 9.2 for any rod reading, the
rod plumbing error is approximated as
eLS ⫽ r ⫺ r⬘ ⫽
d2
2r
(9.11)
where d is the linear amount that the rod is out of plumb at the location of
the rod reading, r. The size of d is dependent on the rod level bubble centering
error and the reading location. If the rod bubble is out of level by , then d
is
d ⫽ r sin 
(9.12)
Substituting Equation (9.12) into (9.11) gives
eLS ⫽
r
sin2
2
(9.13)
Example 9.3 Assume that a rod level bubble is within Ⳳ5⬘ of level and the
rod reading is at 4 m. What is the estimated error in the rod reading?
SOLUTION
eLS ⫽
4 2
sin (5⬘) ⫽ 0.004 mm
2
150
ERROR PROPAGATION IN ELEVATION DETERMINATION
Since the rod plumbing error occurs on every sighting, backsight errors
will tend to cancel foresight errors. Thus, with precise leveling techniques,
the combined effect of this error can be written as
e⫽
冉
冊 冉
冊
r1 sin2 r2 sin2
r sin2 r4 sin2
⫺
⫹ 3
⫺
⫹䡠䡠䡠
2
2
2
2
(9.14)
Grouping like terms in Equation (9.14) yields
e ⫽ –12 sin2 (r1 ⫺ r2 ⫹ r3 ⫺ r4 ⫹ 䡠 䡠 䡠)
(9.15)
Recognizing that the quantity in parentheses in Equation (9.15) is the elevation difference for the leveling line yields
eLS ⫽
⌬Elev
2
sin2
(9.16)
Example 9.4 If a level rod is maintained to within Ⳳ5⬘ of level and the
elevation difference is 22.865 m, the estimated error in the final elevation is
eLS
22.865 2
sin (5⬘) ⫽ 0.02 mm
2
The rod plumbing error can be practically eliminated by carefully centering
the bubble of a well-adjusted rod level. It is generally small, as the example
illustrates, and thus will be ignored in subsequent computations.
9.3.4
Estimated Errors in Differential Leveling
From the preceding discussion, the major random error sources in differential
leveling are caused by random errors in rod readings and instrument leveling.
Furthermore, in Equation (9.9), the collimation error is considered to be systematic and is effectively negated by balancing the backsight and foresight
distances. However, no matter what method is used to observe the lengths of
the sight distances, some random error in these lengths will be present. This
causes random errors in the elevation differences, due to the effects of Earth
curvature, refraction, and instrumental collimation errors. Equation (6.16) can
be applied to Equation (9.9) to model the effects of the random errors in rod
readings, leveling, and sighting lengths. The following partial derivatives are
needed:
9.3
⭸⌬h
⭸⌬h
⫽
⫽1
⭸r1
⭸r2
冉
RANDOM ERRORS IN DIFFERENTIAL LEVELING
⭸⌬h
⫽ ⫺D1
⭸␣1
冊
冉
⭸⌬h
⫽ ⫺D2
⭸␣2
151
(9.17)
冊
⭸⌬h
CR(D1) ⭸⌬h
CR(D2)
⫽⫺ ␣⫹
⫽⫺ ␣⫹
⭸D1
500,000 ⭸D2
500,000
By substituting Equations (9.17) into (6.16) with their corresponding estimated standard errors, the standard error in a single elevation difference can
be estimated as
冪
(D1r / D)2 ⫹ (D2r / D)2 ⫹ (⫺D1␣1)2 ⫹ (⫺D2␣2)2
⌬h ⫽
冋冉
冊 册 冋冉
CR(D1)
⫹ ⫺ ␣⫹
500,000 D1
2
冊 册
CR(D2)
␣⫹
500,000 D2
⫹
2
(9.18)
where r / D is the estimated error in a rod reading, ␣1 and ␣2 the estimated
collimation errors in the backsight and foresight, respectively, and D1 and
D2 the errors estimated in the sight lengths D1 and D2, respectively.
In normal differential leveling procedures, D1 ⫽ D2 ⫽ D. Also, the estimated standard errors in the sight distances are equal: D1 ⫽ D2 ⫽ D. Furthermore, the estimated collimation error for the backsight and foresight can
be assumed equal: ␣1 ⫽ ␣2 ⫽ ␣. Thus, Equation (9.18) simplifies to
⌬h ⫽
冪
冉
2D2 (2r / D ⫹ 2␣) ⫹ 2D2 ␣ ⫹
冊
CR(D)
500,000
2
(9.19)
Equation (9.19) is appropriate for a single elevation difference when the
sight distances are approximately equal. In general, if sight distances are kept
equal for N instrument setups, the total estimated error in an elevation difference is
⌬h ⫽
冪
冉
2ND2 (r2/ D ⫹ 2␣) ⫹ 2N2D ␣ ⫹
冊
CR(D)
500,000
2
(9.20)
Since the error estimated in the elevation difference due to Earth curvature
and refraction, and the actual collimation error, ␣, are small, the last term is
ignored. Thus, the final equation for the standard error estimated in differential leveling is
⌬h ⫽ D兹2N(r2/ D ⫹ 2␣)
(9.21)
152
ERROR PROPAGATION IN ELEVATION DETERMINATION
Example 9.5 A level line is run from benchmark A to benchmark B. The
standard error estimated in rod readings is Ⳳ0.01 mm/m. The instrument is
maintained to within Ⳳ2.0ⴖ of level. A collimation test shows that the instrument is within 4 mm per 100 m. Fifty-meter sight distances are maintained
within an uncertainty of Ⳳ2 m. The total line length from A to B is 1000 m.
What is the estimated error in the elevation difference between A and B? If
A had an elevation of 212.345 Ⳳ 0.005 m, what is the estimated error in the
computed elevation of B?
SOLUTION The total number of setups in this problem is 1000/(2 ⫻ 50)
⫽ 10 setups. Substituting the appropriate values into Equation (8.20) yields
⌬h ⫽
冪
2(10)502
冋冉 冊 冉 冊 册
0.01
1000
2
2.0ⴖ
⫹
2
⫹ 2(10)22
冉
冊
0.004 0.0675(50)
⫹
100
500,000
2
⫽ 兹0.0031 ⫹ 0.0004 ⫽ 0.0031 m ⫽ Ⳳ3.1 mm
2
2
From an analysis of the individual error components in the equation above,
it is seen that the error caused by the errors in the sight distances is negligible
for all but the most precise leveling. Thus, like the error due to rod bubble
centering, this error can be ignored in all but the most precise work. Thus,
the simpler Equation (9.21) can be used to solve the problem:
冪
⌬h ⫽ 50 2 ⫻ 10
冋冉 冊 冉 冊 册
0.01
1000
2
⫹
2.0ⴖ
2
⫽ 兹0.0031 ⫽ Ⳳ0.0031 m ⫽ Ⳳ3.1 mm
2
The estimated error in the elevation of B is found by applying Equation (6.18)
as
2
ElevB ⫽ 兹2ElevA ⫹ ⌬Elev
⫽ 兹52 ⫹ 3.12 ⫽ Ⳳ5.9 mm
9.4
ERROR PROPAGATION IN TRIGONOMETRIC LEVELING
With the introduction of total station instruments, it is becoming increasingly
convenient to observe elevation differences using trigonometric methods.
However, in this procedure, because sight distances cannot be balanced, it is
important that the systematic effects of Earth curvature and refraction, and
inclination in the instrument’s line of sight (collimation error), be removed.
9.4
153
ERROR PROPAGATION IN TRIGONOMETRIC LEVELING
Figure 9.3 Determination of elevation difference by trigonometric leveling.
From Figure 9.3, the corrected elevation difference, ⌬h, between two points
is
⌬h ⫽ hi ⫹ S sin v ⫹ hCR ⫺ hr
(9.22)
Equation (9.22) for zenith-angle reading instruments is
⌬h ⫽ hi ⫹ S cos v ⫹ hCR ⫺ hr
(9.23)
where hi is the instrument height above ground, S is the slope distance between the two points, v the vertical angle between the instrument and the
prism, z the zenith angle between the instrument and the prism, hCR the Earth
curvature and refraction correction given in Equation (9.4), and hr the rod
reading. Substituting the curvature and refraction formula into Equation (9.23)
yields
冉 冊
⌬h ⫽ hi ⫹ cos z ⫹ CR
S sin z
1000
2
⫺ hr
(9.24)
In developing an error propagation formula for Equation (9.24), not only
must errors relating to the height of instrument and prism be considered, but
also, errors in leveling, pointing, reading, and slope distances as discussed in
Chapter 6. Applying Equation (6.16) to Equation (9.24), the following partial
derivatives apply:
154
ERROR PROPAGATION IN ELEVATION DETERMINATION
⭸⌬h
⫽1
⭸hi
⭸⌬h
⫽ ⫺1
⭸hr
⭸⌬h
CR(S) sin2z
⫽ cos z ⫹
⭸S
500,000
⭸⌬h
CR(S 2)sin z cos z
⫽
⫺ S sin z
⭸z
500,000
Entering the partial derivatives and the standard errors of the observations
into Equation (6.16), the total error in trigonometric leveling is
冪 冋冉
2hi ⫹ 2hr ⫹
⌬h ⫽
⫹
冋冉
cos z ⫹
冊 册
冊 册
2
CR(S) sin2z
S
500,000
CR(S 2) sin z cos z
Z
⫺ S sin z
500,000
2
(9.25)
where z is the zenith angle, CR is 0.0675 if units of meters are used or 0.0206
if units of feet are used, S is the slope distance, and is the radians-to-seconds
conversion of 206,264.8ⴖ /rad.
In Equation (9.25), errors from several sources make up the estimated error
in the zenith angle. These include the operator’s ability to point and read the
instrument, the accuracy of the vertical compensator or the operator’s ability
to center the vertical circle bubble, and the sensitivity of the compensator or
vertical circle bubble. For best results, zenith angles should be observed using
both faces of the instrument and an average taken. Using Equations (7.4) and
(7.6), the estimated error in a zenith angle that is observed in both positions
(face I and face II) with a theodolite is
z ⫽
冪
22r ⫹ 22p ⫹ 2B2
N
(9.26a)
where r is the error in reading the circle, p the error in pointing, B the
error in the vertical compensator or in leveling the vertical circle bubble, and
N the number of face-left and face-right observations of the zenith angle. For
digital theodolites or total stations, the appropriate formula is
9.4
ERROR PROPAGATION IN TRIGONOMETRIC LEVELING
z ⫽
冪
22DIN ⫹ 22B
N
155
(9.26b)
where DIN is the DIN 18723 value for the instrument and all other values
are as above.
Notice that if only a single zenith-angle observation is made (i.e., it is
observed only in face I), its estimated error is simply
z ⫽ 兹r2 ⫹ p2 ⫹ B2
(9.27a)
For the digital theodolite and total station, the estimated error for a single
observation is
2
z ⫽ 兹2DIN
⫹ B2
(9.27b)
Similarly, the estimated error in the slope distance, S, is computed using
Equation (6.36).
Example 9.6 A total station instrument has a vertical compensator accurate
to within Ⳳ0.3ⴖ, a digital reading accuracy of Ⳳ5ⴖ, and a distance accuracy
of Ⳳ(5 mm ⫹ 5 ppm). The slope distance observed is 1256.78 ft. It is estimated that the instrument is set to within Ⳳ 0.005 ft of the station, and the
target is set to within Ⳳ0.01 ft. The height of the instrument is 5.12 Ⳳ 0.01
ft, and the prism height is 6.72 Ⳳ 0.01 ft. The zenith angle is observed in
only one position and recorded as 88⬚13⬘15ⴖ. What are the corrected elevation
difference and its estimated error?
SOLUTION Using Equation (9.24), the corrected elevation difference is
⌬h ⫽ 5.12 ⫹ 1256.78 cos(88⬚13⬘15⬘) ⫹ 0.0206
冉
冊
1256.78 sin(88⬚13⬘15⬘)
1000
2
⫺ 6.72
⫽ 37.39 ft
With Equation (9.27b), the zenith angle error is estimated as z ⫽
兹2(5)2 ⫹ 0.32 ⫽ Ⳳ7.1ⴖ. From Equation (7.24) and converting 5 mm to
0.0164 ft, the estimated error in the distance is
156
ERROR PROPAGATION IN ELEVATION DETERMINATION
S ⫽
冪
0.0052 ⫹ 0.012 ⫹ 0.01642 ⫹
冉
冊
5
1256.78
1,000,000
2
⫽ Ⳳ0.021 ft
Substituting the values into Equation (9.25), the estimated error in the elevation difference is
⌬h ⫽
冉
冪
0.012 ⫹ 0.012 ⫹ (0.031 ⫻ 0.021)2 ⫹ ⫺1256.172
冊
7.1ⴖ
2
⫽ 兹0.01 ⫹ 0.01 ⫹ 0.00065 ⫹ 0.043 ⫽ Ⳳ0.045 ft
2
2
2
2
Note in this example that the estimated error in the elevation difference
caused by the distance error is negligible (Ⳳ0.00065 ft), whereas the error in
the zenith angle is the largest (Ⳳ0.043 ft). Furthermore, since the vertical
angle is not observed with both faces of the instrument, it is possible that
uncompensated systematic errors are present in the final value computed. For
example, assume that a 10ⴖ indexing error existed on the vertical circle. If
the observations are taken using both faces of the instrument, the effects of
this error are removed. However, by making only a face I observation, the
systematic error due to the vertical indexing error is
1256.78 sin(10ⴖ) ⫽ 0.061 ft
The uncompensated systematic error in the final value is considerably larger
than the error estimated for the observation. One should always account for
systematic errors by using proper field procedures. Failure to do so can only
lead to poor results. In trigonometric leveling, a minimum of a face I and
face II reading should always be taken.
PROBLEMS
9.1
A collimation error of 0.00005 ft/ft is used in leveling a line that has
a sum of 1425 ft for the backsight distances and only 632 ft for the
foresight distances. If the elevation difference observed is ⫺15.84 ft,
what is the elevation difference corrected for the collimation error?
9.2
Repeat Problem 9.1 for a collimation error of 0.005 mm/m, a sum of
the backsight distances of 823 m, a sum of the foresight distance of
1206 m, and an observed elevation difference of 23.475 m.
9.3
To expedite going down a hill, the backsight distances were consistently held to 50 ft while the foresight distances were held to 200 ft.
PROBLEMS
157
There were 33 setups. If the elevation difference observed was
⫺306.87 ft, what is the elevation difference corrected for Earth curvature and refraction?
9.4
Repeat Problem 9.3 for backsight distances of 20 m, foresight distances
of 60 m, 46 setups, and an elevation difference of ⫺119.603 m.
9.5
How far from vertical must an 8-ft level rod be to create an error of
0.01 ft with a reading at 5.00 ft?
9.6
Repeat Problem 9.5 for a reading at 15.00 ft on a 25-ft rod.
9.7
Repeat Problem 9.5 for a 1-mm error with a reading at 1.330 m on a
2-m rod.
9.8
A line of three-wire differential levels goes from benchmark Gloria to
benchmark Carey. The length of the line was determined to be 2097
m. The instrument had a stated compensator accuracy of Ⳳ1.4 mm/
km. The instrument–rod combination had an estimated reading error
of Ⳳ0.4 mm per 40 m. Sight distances were kept to approximately 50
Ⳳ 5 m. If the observed difference in elevation is 15.601 m, what is
the estimated error in the final elevation difference?
9.9
If in Problem 9.8, benchmark Gloria had a fixed elevation of 231.071
m and Carey had a fixed elevation of 246.660 m, did the job meet
acceptable closure limits at a 95% level of confidence? Justify your
answer statistically.
9.10 Repeat Problem 9.8 for a compensator accuracy of Ⳳ1.2 mm/km and
an estimated error in reading the rod of 0.4 mm per 100 m.
9.11 Using the data in Problem 9.9, did the leveling circuit meet acceptable
closures at a 95% level of confidence? Justify your response statistically.
9.12 An elevation must be established on a benchmark on an island that is
2536.98 ft from the nearest benchmark on the island’s shore. The surveyor decides to use a total station that has a stated distance measuring
accuracy of Ⳳ(3 mm ⫹ 3 ppm) and a vertical compensator accurate
to within Ⳳ0.4ⴖ. The height of instrument was 5.37 ft with an estimated
error of Ⳳ0.05 ft. The prism height was 6.00 ft with an estimated error
of Ⳳ0.02 ft. The single zenith angle is read as 87⬚05⬘32ⴖ. The estimated
errors in instrument and target centering are Ⳳ0.003 ft. If the elevation
of the occupied benchmark is 632.27 ft, what is the corrected benchmark elevation on the island? (Assume that the instrument does not
correct for Earth curvature and refraction.)
9.13 In Problem 9.12, what is the estimated error in the computed benchmark elevation if the instrument has a DIN 18723 stated accuracy of
Ⳳ3ⴖ
158
ERROR PROPAGATION IN ELEVATION DETERMINATION
9.14 After completing the job in Problem 9.12, the surveyor discovered that
the instrument had a vertical indexing error that caused the sight line
to be inclined by 15ⴖ.
(a) How much error would be created in the elevation of the island
benchmark if the indexing error were ignored?
(b) What is the corrected elevation?
9.15 A tilting level is used to run a set of precise levels to a construction
project from benchmark DAM, whose elevation is 101.865 m. The line
is run along a road that goes up a steep incline. To expedite the job,
backsight distances are kept to 50 Ⳳ 1 m and foresight distances are
20 Ⳳ 1 m. The total length of differential levels is 4410 m. The elevation difference observed is 13.634 m. What is the corrected project
elevation if the instrument has a sight line that declines at the rate of
0.3 mm per 50 m?
9.16 In Problem 9.15, the instrument’s level is centered to within Ⳳ0.4ⴖ for
each sight and the rod is read to Ⳳ1.2 mm per 50 m.
(a) What is the estimated error in the elevation difference?
(b) What is the 95% error in the final established benchmark?
9.17 Which method of leveling presented in this chapter offers the most
precision? Defend your answer statistically.
Programming Problems
9.18 Create a computational package that will compute a corrected elevation
difference and its estimated error using the method of differential leveling. Use this package to solve Problem 9.15.
9.19 Create a computational package that will compute a corrected elevation
difference and its estimated error using the method of trigonometric
leveling. Use this package to solve Problem 9.12.
CHAPTER 10
WEIGHTS OF OBSERVATIONS
10.1
INTRODUCTION
When surveying data are collected, they must usually conform to a given set
of geometric conditions, and when they do not, the measurements are adjusted
to force that geometric closure. For a set of uncorrelated observations, a
measurement with high precision, as indicated by a small variance, implies a
good observation, and in the adjustment it should receive a relatively small
portion of the overall correction. Conversely, a measurement with lower precision, as indicated by a larger variance, implies an observation with a larger
error, and should receive a larger portion of the correction.
The weight of an observation is a measure of its relative worth compared
to other measurements. Weights are used to control the sizes of corrections
applied to measurements in an adjustment. The more precise an observation,
the higher its weight; in other words, the smaller the variance, the higher the
weight. From this analysis it can be stated intuitively that weights are inversely proportional to variances. Thus, it also follows that correction sizes
should be inversely proportional to weights.
In situations where measurements are correlated, weights are related to the
inverse of the covariance matrix, ⌺. As discussed in Chapter 6, the elements
of this matrix are variances and covariances. Since weights are relative, variances and covariances are often replaced by cofactors. A cofactor is related
to its covariance by the equation
qij ⫽
ij
02
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
(10.1)
159
160
WEIGHTS OF OBSERVATIONS
where qij is the cofactor of the ijth measurement, ij the covariance of the ijth
measurement, and 20 the reference variance, a value that can be used for
scaling. Equation (10.1) can be expressed in matrix notation as
Q⫽
1
20
⌺
(10.2)
where Q is defined as the cofactor matrix. The structure and individual elements of the ⌺ matrix are
⌺⫽
2x1
x2x1
⯗
xnx1
x1x2
2x2
⯗
xnx2
䡠䡠䡠
x1xn
x2xn
⯗
2xn
From the discussion above, the weight matrix W is
W ⫽ Q⫺1 ⫽ 02 ⌺⫺1
(10.3)
For uncorrelated measurements, the covariances are equal to zero (i.e., all
xixj ⫽ 0) and the matrix ⌺ is diagonal. Thus, Q is also a diagonal matrix
with elements equal to 2xi / 20. The inverse of a diagonal matrix is also a
diagonal matrix, with its elements being the reciprocals of the original diagonals, and therefore Equation (10.3) becomes
W⫽
20
2x1
0
0
20
2x2
0
0
⫽ 02 ⌺⫺1
(10.4)
䡠䡠䡠
0
0
䡠䡠䡠
20
2xn
0
From Equation (10.4), any independent measurement with variance equal to
2i has a weight of
wi ⫽
20
i2
(10.5)
If the ith observation has a weight wi ⫽ 1, then 20 ⫽ 2i , or 20 ⫽ 1. Thus,
is often called the variance of an observation of unit weight, shortened to
2
0
10.2
WEIGHTED MEAN
161
variance of unit weight or simply unit variance. Its square root is called the
standard deviation of unit weight. If 20 is set equal to 1 in Equation (10.5),
then
wi ⫽
1
i2
(10.6)
Note in Equation (10.6) that as stated earlier, the weight of an observation is
inversely proportional to its variance.
With correlated observations, it is possible to have a covariance matrix, ⌺,
and a cofactor matrix, Q, but not a weight matrix. This occurs when the
cofactor matrix is singular, and thus an inverse for Q does not exist. Most
situations in surveying involve uncorrelated observations. For the remainder
of this chapter, only the uncorrelated case with variance of unit weight is
considered.
10.2
WEIGHTED MEAN
If two measurements are taken of a quantity and the first is twice as good as
the second, their relative worth can be expressed by giving the first measurement a weight of 2 and the second a weight of 1. A simple adjustment
involving these two measurements would be to compute the mean value. In
this calculation, the observation of weight 2 could be added twice, and the
observation of weight 1 added once. As an illustration, suppose that a distance
is measured with a tape to be 151.9 ft, and the same distance is measured
with an EDM instrument as 152.5 ft. Assume that experience indicates that
the electronically measured distance is twice as good as the taped distance,
and accordingly, the taped distance is given a weight of 1 and the electronically measured distance is given a weight of 2. Then one method of computing the mean from these observations is
M⫽
151.9 ⫹ 152.5 ⫹ 152.5
⫽ 152.3
3
As an alternative, the calculation above can be rewritten as
M⫽
1(151.9) ⫹ 2(152.5)
⫹ 152.3
1⫹2
Note that the weights of 1 and 2 were entered directly into the second
computation and that the result of this calculation is the same as the first.
Note also that the computed mean tends to be closer to the measured value
162
WEIGHTS OF OBSERVATIONS
having the higher weight (i.e., 152.3 is closer to 152.5 than it is to 151.9). A
mean value computed from weighted observations is called the weighted
mean.
To develop a general expression for computing the weighted mean, suppose
that we have m independent uncorrelated observations (z1, z2, . . . , zm) for a
quantity z and that each observation has standard deviation . Then the mean
of the observations is
z⫽
mi⫽1 zi
(10.7)
m
If these m observations were now separated into two sets, one of size ma and
the other mb such that ma ⫹ mb ⫽ m, the means for these two sets would be
za ⫽
zb ⫽
ma
i⫽1
zi
(10.8)
ma
mi⫽ma⫹1 zi
(10.9)
mb
The mean z is found by combining the means of these two sets as
z⫽
mi⫽1a zi ⫹ mi⫽ma⫹1 zi
m
⫽
mi⫽1a zi ⫹ mi⫽ma⫹1 zi
ma ⫹ mb
(10.10)
But from Equations (10.8) and (10.9),
a
zama ⫽ mi⫽1
zi and
m
zbmb ⫽ i⫽m
zi
a⫹1
(10.11)
Thus,
z⫽
zama ⫹ zbmb
ma ⫹ mb
(10.12)
Note the correspondence between Equation (10.12) and the second equation
used to compute the weighted mean in the simple illustration given earlier.
By intuitive comparison it should be clear that ma and mb correspond to
weights that could be symbolized as wa and wb, respectively. Thus, Equation
(10.12) can be written as
z⫽
waza ⫹ wbzb wz
⫽
wa ⫹ wb
w
(10.13)
Equation (10.13) is used in calculating the weighted mean for a group of
uncorrelated observations having unequal weights. In Chapter 11 it will be
10.3
163
RELATION BETWEEN WEIGHTS AND STANDARD ERRORS
shown that the weighted mean is the most probable value for a set of weighted
observations.
Example 10.1 As a simple example of computing a weighted mean using
Equation (10.13), suppose that a distance d is measured three times, with the
following results: 92.61 with weight 3, 92.60 with weight 2, and 92.62 with
weight 1. Calculate the weighted mean.
d⫽
3(92.61) ⫹ 2(92.60) ⫹ 1(92.62)
⫽ 92.608
3⫹2⫹1
Note that if weight had been neglected, the simple mean would have been
92.61.
10.3
RELATION BETWEEN WEIGHTS AND STANDARD ERRORS
By applying the special law of propagation of variances [Equation (6.16)] to
Equation (10.8), the variance za in Equation (10.8) is
2za ⫽
冉 冊
⭸za
⭸z1
2
2 ⫹
冉 冊
⭸za
⭸z2
2
2 ⫹ 䡠 䡠 䡠 ⫹
冉 冊
2
⭸za
⭸zma
2
(10.14)
Substituting partial derivatives with respect to the measurements into Equation
(10.14) yields
冉冊 冉冊
2
2za ⫽
2
冉冊
2
1
1
1
2 ⫹
2 ⫹ 䡠 䡠 䡠 ⫹
2
ma
ma
ma
Thus,
冉冊
2
1
1 2
⫽ ma
2 ⫽
ma
ma
2
za
(10.15)
Using a similar procedure, the variance of zb is
2zb ⫽
1 2
mb
(10.16)
In Equations (10.15) and (10.16), is a constant, and the weights of za
and zb were established as ma and mb, respectively, from Equation (10.13).
Since the weights are relative, from Equations (10.15) and (10.16),
164
WEIGHTS OF OBSERVATIONS
wa ⫽
1
and wb ⫽
z2a
1
(10.17)
z2b
Conclusion: With uncorrelated observations, the weights of the observations
are inversely proportional to their variances.
10.4
10.4.1
STATISTICS OF WEIGHTED OBSERVATIONS
Standard Deviation
By definition, an observation is said to have a weight w when its precision is
equal to that of the mean of w observations of unit weight. Let 0 be the
standard error of an observation of weight 1, or unit weight. If y1, y2, . . . ,
yn are observations having standard errors 1, 2, . . . , n and weights w1,
w2, . . . , wn, then, by Equation (10.5),
0
1 ⫽
兹w1
, 2 ⫽
0
兹w2
, . . . , n ⫽
0
兹wn
(10.18)
In Section 2.7, the standard error for a group of observations of equal
weight was defined as
⫽
ni⫽1 ε2i
n
Now, in the case where the observations are not equal in weight, the equation
above becomes
⫽
w1ε12 ⫹ w2ε22 ⫹ 䡠 䡠 䡠 wnε2n
⫽
n
n
i⫽1
wi ε2i
n
(10.19)
When modified for the standard deviation in Equation (2.7), it is
S⫽
w1v12 ⫹ w2v22 ⫹ 䡠 䡠 䡠 ⫹ wnv2n
⫽
n⫺1
n
i⫽1
wiv2i
n⫺1
(10.20)
10.4.2 Standard Error of Weight w and Standard Error of the
Weighted Mean
The relationship between standard error and standard error of weight w was
given in Equation (10.18). Combining this with Equation (10.19) and drop-
10.5
WEIGHTS IN ANGLE OBSERVATIONS
165
ping the summation limits, equations for standard errors of weight w are
obtained in terms of 0 as follows:
1 ⫽
2 ⫽
0
⫽
w1
0
⫽
w2
wε2 1
w1
n
wε2 1
w2
n
⫽
wε2
nw
1
⫽
wε2
nw
(10.21)
2
⯗
n ⫽
0
⫽
wn
wε2 1
wn
n
⫽
wε2
nw
n
Similarly, standard deviations of weight w can be expressed as
S1 ⫽
wv2
, S2 ⫽
w1(n ⫺ 1)
wv2
, . . . , Sn ⫽
w2(n ⫺ 1)
wv2
w (n ⫺1)
(10.22)
n
Finally, the reference standard error of the weighted mean is calculated as
wε2
nw
M ⫽
(10.23)
and the standard deviation of the weighted mean is
M ⫽
10.5
wv2
(n ⫺1) w
(10.24)
WEIGHTS IN ANGLE OBSERVATIONS
Suppose that the three angles ␣1, ␣2, and ␣3 in a plane triangle are observed
n1, n2, and n3 times, respectively, with the same instrument under the same
conditions. What are the relative weights of the angles?
To analyze the relationship between weights and the number of times an
angle is turned, let S be the standard deviation of a single angle observation.
The means of the three angles are
166
WEIGHTS OF OBSERVATIONS
␣1 ⫽
␣1
n1
␣2 ⫽
␣2
n2
␣3 ⫽
␣3
n3
The variances of the means, as obtained by applying Equation (6.16), are
S 2␣1 ⫽
1 2
S
n1
S 2␣2 ⫽
1 2
S
n2
S 2␣3 ⫽
1 2
S
n3
Again, since the weights of the observations are inversely proportional to the
variances and relative, the weights of the three angles are
w1 ⫽
1
n
⫽ 12
2
S ␣1 S
w2 ⫽
1
n
⫽ 22
2
S ␣2 S
w3 ⫽
1
n
⫽ 32
2
S ␣3 S
In the expressions above, S is a constant term in each of the weights, and
because the weights are relative, it can be dropped. Thus, the weights of the
angles are w1 ⫽ n1, w2 ⫽ n2, and w3 ⫽ n3.
In summary, it has been shown that when all conditions in angle observation are equal except for the number of turnings, angle weights are proportional to the number of times the angles are turned.
10.6
WEIGHTS IN DIFFERENTIAL LEVELING
Suppose that for the leveling network shown in Figure 10.1, the lengths of
lines 1, 2, and 3 are 2, 3, and 4 miles, respectively. For these varying lengths
of lines, it can be expected that the errors in their elevation differences will
vary, and thus the weights assigned to the elevation differences should also
be varied. What relative weights should be used for these lines?
To analyze the relationship of weights and level line lengths, recall from
Equation (9.20) that the variance in ⌬h is
2⌬h ⫽ D2[2N(2r / D ⫹ 2␣)]
Figure 10.1 Differential leveling network.
(a)
10.7
PRACTICAL EXAMPLES
167
where D is the length of the individual sights, N the number of setups, r / D
the estimated error in a rod reading, and ␣ the estimated collimation error
for each sight. Let li be the length of the ith course between benchmarks;
then
N⫽
li
2D
(b)
Substituting Equation (b) into Equation (a) yields
2⌬h ⫽ li D(2r / D ⫹ 2␣)
(c)
However, D, r / D, and ␣ are constants, and thus by letting k ⫽ D(2r / D ⫹
2␣), Equation (c) becomes
2⌬h ⫽ li k
(d)
For this example, it can be said that the weights are
w1 ⫽
1
l1k
w2 ⫽
1
l2k
w3 ⫽
1
l3k
(e)
Now since k is a constant and weights are relative, Equation (e) can be simplified to
w1 ⫽
1
l1
w2 ⫽
1
l2
w3 ⫽
1
l3
In summary, it has been shown that weights of differential leveling lines
are inversely proportional to their lengths, and since any course length is
proportional to its number of setups, weights are also inversely proportional
to the number of setups.
10.7
PRACTICAL EXAMPLES
Example 10.2 Suppose that the angles in an equilateral triangle ABC were
observed by the same operator using the same instrument, but the number of
repetitions for each angle varied. The results were A ⫽ 45⬚15⬘25ⴖ, n ⫽ 4; B
⫽ 83⬚37⬘22ⴖ, n ⫽ 8; and 51⬚07⬘39ⴖ, n ⫽ 6. Adjust the angles.
SOLUTION Weights proportional to the number of repetitions are assigned
and corrections are made in inverse proportion to those weights. The sum of
the three angles is 180⬚00⬘26ⴖ, and thus the misclosure that must be adjusted
is 26ⴖ. The correction process is demonstrated in Table 10.1.
168
WEIGHTS OF OBSERVATIONS
TABLE 10.1 Adjustment of Example 10.2
Angle
n
(Weight)
Correction
Factor
Correction
Corrected
Angle
A
B
C
4
8
6
(1 / 4) ⫻ 24 ⫽ 6
(1 / 8) ⫻ 24 ⫽ 3
(1 / 6) ⫻ 24 ⫽ 4
13
(6 / 13) ⫻ 26 ⫽ 12ⴖ
(3 / 13) ⫻ 26 ⫽ 06ⴖ
(4 / 13) ⫻ 26 ⫽ 08ⴖ
26ⴖ
45⬚15⬘13ⴖ
83⬚37⬘16ⴖ
51⬚07⬘31ⴖ
180⬚00⬘00ⴖ
Note that a multiplier of 24 was used for convenience to avoid fractions
in computing correction factors. Because weights are relative, this did not
alter the adjustment. Note also that two computational checks are secured in
the solution above; the sum of the individual corrections totaled 26ⴖ, and the
sum of the corrected angles totaled 180⬚00⬘00ⴖ.
Example 10.3 In the leveling network of Figure 10.1, recall that the lengths
of lines 1, 2, and 3 were 2, 3, and 4 miles, respectively. If the observed
elevation differences in lines 1, 2, and 3 were ⫹21.20 ft, ⫹21.23 ft, and
⫹21.29 ft, respectively, find the weighted mean for the elevation difference
and the adjusted elevation of BMX. (Note: All level lines were run from BMA
to BMX.)
SOLUTION The weights of lines 1, 2, and 3 are 1/2, 1/3, and 1/4, respectively. Again since weights are relative, these weights can arbitrarily be
multiplied by 12 to obtain weights of 6, 4, and 3, respectively. Applying
Equation (10.13), the weighted mean is
mean ⌬Elev ⫽
6(21.20) ⫹ 4(21.23) ⫹ 3(21.29)
⫽ ⫹21.23
6⫹4⫹3
Thus, the elevation of BMX ⫽ 100.00 ⫹ 21.23 ⫽ 123.23 ft. Note that if the
weights had been neglected, the simple average would have given a mean
elevation difference of ⫹21.24.
Example 10.4 A distance is measured as 625.79 ft using a cloth tape and a
given weight of 1; it is measured again as 625.71 ft using a steel tape and
assigned a weight of 2; and finally, it is measured a third time as 625.69 ft
with an EDM instrument and given a weight of 4. Calculate the most probable
value of the length (weighted mean), and find the standard deviation of the
weighted mean.
10.7
PRACTICAL EXAMPLES
169
SOLUTION By Equation (10.13), the weighted mean is
M⫽
1(625.79) ⫹ 2(623.71) ⫹ 4(625.69)
⫽ 625.71 ft
1⫹2⫹4
By Equation (10.24), the standard deviation of the weighted mean is
SM ⫽
wv2
⫽
(n ⫺1) w
⫽ Ⳳ0.024 ft
0.0080
(2)7
where
v1 ⫽ 625.71 ⫺ 625.79 ⫽ ⫺0.08
w1v12 ⫽ 1(⫺0.08)2 ⫽ 0.0064
v2 ⫽ 625.71 ⫺ 625.71 ⫽
w2v22 ⫽ 2(0.00)2
0.00
v3 ⫽ 625.71 ⫺ 625.69 ⫽ ⫹0.02
⫽ 0.0000
w3v23 ⫽ 4(⫹0.02)2 ⫽ 0.0016
0.0080
Example 10.5 In leveling from benchmark A to benchmark B, four different
routes of varying length are taken. The data of Table 10.2 are obtained. (Note
that the weights were computed as 18/l for computational convenience only.)
Calculate the most probable elevation difference (weighted mean), the standard deviation of unit weight, the standard deviation of the weighted mean,
and the standard deviations of the weighted observations.
SOLUTION By Equation (10.13), the weighted mean for elevation difference is
M⫽
18(25.35) ⫹ 9(25.41) ⫹ 6(25.38) ⫹ 3(25.30)
⫽ ⫹25.366 ft
18 ⫹ 9 ⫹ 6 ⫹ 3
TABLE 10.2 Route Data for Example 10.5
Route
Length (miles)
⌬Elev
w
1
2
3
4
1
2
3
6
⫹25.35
⫹25.41
⫹25.38
⫹25.30
18
9
6
3
170
WEIGHTS OF OBSERVATIONS
TABLE 10.3 Data for Standard Deviations in Example 10.5
Route
w
v
v2
wv2
1
2
3
4
18
9
6
3
⫹0.016
⫺0.044
⫺0.014
⫹0.066
0.0002
0.0019
0.0002
0.0043
0.0045
0.0174
0.0012
0.0130
0.0361
The arithmetic mean for this set of observations is 25.335, but the weighted
mean is 25.366. To find the standard deviations for the weighted observations,
the data in Table 10.3 are first created.
By Equation (10.20), the standard deviation of unit weight is
S0 ⫽
⫽ Ⳳ0.11 ft
冪0.0361
3
By Equation (10.24), the standard deviation of the weighted mean is
SM ⫽
⫽ Ⳳ0.018 ft
冪0.0361
36(3)
By Equation (10.22), the standard deviations for the weighted observations
are
S1 ⫽
⫽ Ⳳ0.026 ft
冪0.0361
18(3)
S2 ⫽
⫽ Ⳳ0.037 ft
冪0.0361
9(3)
S3 ⫽
⫽ Ⳳ0.045 ft
冪0.0361
6(3)
S4 ⫽
⫽ Ⳳ0.063 ft
冪0.0361
3(3)
PROBLEMS
10.1
An angle was measured as 49⬚27⬘30ⴖ using an engineer’s transit and
had a standard deviation of Ⳳ30ⴖ. It was measured again using a
repeating optical theodolite as 49⬚27⬘24ⴖ with a standard deviation of
Ⳳ10ⴖ. This angle was measured a third time with a directional theodolite as 49⬚27⬘22ⴖ with a standard deviation of Ⳳ2ⴖ. Calculate the
weighted mean of the angle and its standard deviation.
PROBLEMS
10.2
171
An angle was measured at four different times, with the following
results:
Day
Angle
S
1
2
3
4
136⬚14⬘34ⴖ
136⬚14⬘36ⴖ
136⬚14⬘28ⴖ
136⬚14⬘26ⴖ
Ⳳ12.2ⴖ
Ⳳ6.7ⴖ
Ⳳ8.9ⴖ
Ⳳ9.5ⴖ
What is the most probable value for the angle and the standard deviation in the mean?
10.3
A distance was measured by pacing as 154 ft with a standard deviation of Ⳳ2.5 ft. It was then observed as 153.86 ft with a steel tape
having a standard deviation of Ⳳ0.05 ft. Finally, it was measured as
153.89 ft with an EDM instrument with a standard deviation of Ⳳ0.02
ft. What is the most probable value for the distance and its standard
deviation?
10.4
A distance was measured by pacing as 267 ft with a standard deviation of Ⳳ3 ft. It was then measured as 268.94 ft with a steel tape
and had a standard deviation of Ⳳ0.05 ft. Finally, it was measured
as 268.99 ft with an EDM. The EDM instrument and reflector setup
standard deviations were Ⳳ0.005 ft and Ⳳ0.01 ft, respectively, and
the manufacturer’s estimated standard deviation for the EDM instrument is Ⳳ(3 mm ⫹ 3 ppm). What is the most probable value for the
distance and the standard deviation of the weighted mean?
10.5
What standard deviation is computed for each weighted observation
of Problem 10.4?
10.6
Compute the standard deviation for the taped observation in Problem
10.4 assuming standard deviations of Ⳳ5⬚F in temperature, Ⳳ3 lb in
pull, Ⳳ0.005 ft in tape length, and Ⳳ0.01 ft in reading and marking
the tape. The temperature at the time of the observation was 78⬚F, the
calibrated tape length was 99.993 ft, and the field tension was recorded as 20 lb. The cross-sectional area of the tape is 0.004 in2, its
modulus of elasticity is 29,000,000 lb/in2, its coefficient of thermal
expansion is 6.45 ⫻ 10⫺5 ⬚F⫺1, and its weight is 2.5 lb. Assume
horizontal taping with full tape lengths for all but the last partial
distance with ends-only support.
10.7
Do Problem 10.3 using the standard deviation information for the
EDM distance in Problem 10.4 and the tape calibration data in Problem 10.6.
172
WEIGHTS OF OBSERVATIONS
10.8
A zenith angle was measured six times with both faces of a total
station. The average direct reading is 88⬚05⬘16ⴖ with a standard deviation of Ⳳ12.8ⴖ. With the reverse face, it was observed as
271⬚54⬘32ⴖ with a standard deviation of Ⳳ9.6ⴖ. What is the most
probable value for the zenith angle in the direct face?
10.9
Three crews level to a benchmark following three different routes.
The lengths of the routes and the observed differences in elevation
are:
Route
⌬Elev (ft)
Length (ft)
1
2
3
14.80
14.87
14.83
3200
4800
3900
What is:
(a) the best value for the difference in elevation?
(b) the standard deviation for the weighted elevation difference?
(c) the standard deviation for the weighted observations?
10.10 Find the standard deviation for Problem 10.9.
CHAPTER 11
PRINCIPLES OF LEAST SQUARES
11.1
INTRODUCTION
In surveying, observations must often satisfy established numerical relationships known as geometric constraints. As examples, in a closed polygon traverse, horizontal angle and distance measurements should conform to the
geometric constraints given in Section 8.4, and in a differential leveling loop,
the elevation differences should sum to a given quantity. However, because
the geometric constraints meet perfectly rarely, if ever, the data are adjusted.
As discussed in earlier chapters, errors in observations conform to the laws
of probability; that is, they follow normal distribution theory. Thus, they
should be adjusted in a manner that follows these mathematical laws. Whereas
the mean has been used extensively throughout history, the earliest works on
least squares started in the late eighteenth century. Its earliest application was
primarily for adjusting celestial observations. Laplace first investigated the
subject and laid its foundation in 1774. The first published article on the
subject, entitled ‘‘Méthode des moindres quarrés’’ (Method of Least Squares),
was written in 1805 by Legendre. However, it is well known that although
Gauss did not publish until 1809, he developed and used the method extensively as a student at the University of Göttingen beginning in 1794 and thus
is given credit for the development of the subject. In this chapter, equations
for performing least squares adjustments are developed and their use is illustrated with several examples.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
173
174
11.2
PRINCIPLES OF LEAST SQUARES
FUNDAMENTAL PRINCIPLE OF LEAST SQUARES
To develop the principle of least squares, a specific case is considered. Suppose that there are n independent equally weighted measurements, z1, z2,
. . . , zn, of the same quantity z, which has a most probable value denoted by
M. By definition,
M ⫺ z1 ⫽ v1
M ⫺ z2 ⫽ v2
(11.1)
⯗
M ⫺ zn ⫽ vn
where the v’s are the residual errors. Note that residuals behave in a manner
similar to errors, and thus they can be used interchangeably in the normal
distribution function given by Equation (3.2). Substituting v for x, there results
ƒx(v) ⫽ y ⫽
1
2
2
2 2
e⫺v / 2 ⫽ Ke⫺h v
2
(11.2)
where h ⫽ 1/ 2 and K ⫽ h/ .
As discussed in Chapter 3, probabilities are represented by areas under the
normal distribution curve. Thus, the individual probabilities for the occurrence
of residuals v1, v2, . . . , vn are obtained by multiplying their respective
ordinates y1, y2, . . . , yn by some infinitesimally small increment of v, ⌬v.
The following probability statements result:
2 2
P1 ⫽ y1 ⌬v ⫽ Ke⫺h v1 ⌬v
2 2
P2 ⫽ y2 ⌬v ⫽ Ke⫺h v2 ⌬v
(11.3)
⯗
2 2
Pn ⫽ yn ⌬v ⫽ Ke⫺h vn ⌬v
From Equation (3.1), the probability of the simultaneous occurrence of all the
residuals v1 through vn is the product of the individual probabilities and thus
2 2
2 2
2 2
P ⫽ (Ke⫺h v1 ⌬v)(Ke⫺h v2 ⌬v) 䡠 䡠 䡠 (Ke⫺h vn ⌬v)
Simplifying Equation (11.4) yields
(11.4)
11.2
FUNDAMENTAL PRINCIPLE OF LEAST SQUARES
2
2
2
2
P ⫽ Kn(⌬v)ne⫺h (v1⫹v2⫹䡠䡠䡠⫹vn)
175
(11.5)
M is the quantity that is to be selected in such a way that it gives the
greatest probability of occurrence, or, stated differently, the value of M that
maximizes the value of P. Figure 11.1 shows a plot of e⫺x versus x. From
this plot it is readily seen that e⫺x is maximized by minimizing x, and thus
in relation to Equation (11.5), the probability P is maximized when the quantity v21 ⫹ v22 ⫹ 䡠 䡠 䡠 ⫹ vn2 is minimized. In other words, to maximize P, the sum
of the squares of the residuals must be minimized. Equation (11.6) expresses
the fundamental principle of least squares:
冘 v ⫽ v ⫹ v ⫹ 䡠 䡠 䡠 ⫹ v ⫽ minimum
2
2
1
2
2
2
n
(11.6)
This condition states: The most probable value (MPV) for a quantity obtained
from repeated observations of equal weight is the value that renders the sum
of the residuals squared a minimum. From calculus, the minimum value of a
function can be found by taking its first derivative and equating the resulting
function with zero. That is, the condition stated in Equation (11.6) is enforced
by taking the first derivative of the function with respect to the unknown
variable M and setting the results equal to zero. Substituting Equation (11.1)
into Equation (11.6) yields
冘 v ⫽ (M ⫺ z ) ⫹ (M ⫺ z ) ⫹ 䡠 䡠 䡠 ⫹ (M ⫺ z )
2
2
1
2
2
2
n
(11.7)
Taking the first derivative of Equation (11.7) with respect to M and setting
the resulting equation equal to zero yields
d(兺 v2)
⫽ 2(M ⫺ z1)(1) ⫹ 2(M ⫺ z2)(1) ⫹ 䡠 䡠 䡠 ⫹ 2(M ⫺ zn)(1) ⫽ 0
dM
Now dividing Equation (11.8) by 2 and simplifying yields
Figure 11.1 Plot of e⫺x.
(11.8)
176
PRINCIPLES OF LEAST SQUARES
M ⫺ z1 ⫹ M ⫺ z2 ⫹ 䡠 䡠 䡠 ⫹ M ⫺ zn ⫽ 0
nM ⫽ z1 ⫹ z2 ⫹ 䡠 䡠 䡠 ⫹ zn
M⫽
z1 ⫹ z2 ⫹ 䡠 䡠 䡠 ⫹ zn
n
(11.9)
In Equation (11.9) the quantity (z1 ⫹ z2 ⫹ 䡠 䡠 䡠 ⫹ zn)/n is the mean of the
values observed. This is proof that when a quantity has been observed independently several times, the MPV is the arithmetic mean.
11.3
FUNDAMENTAL PRINCIPLE OF WEIGHTED LEAST SQUARES
In Section 11.2, the fundamental principle of a least squares adjustment was
developed for observations having equal or unit weights. The more general
case of least squares adjustment assumes that the observations have varying
degrees of precision, and thus varying weights.
Consider a set of measurements z1, z2, . . . , zn having relative weights w1,
w2, . . . , wn and residuals v1, v2, . . . , vn. Denote the weighted MPV as M.
As in Section 11.2, the residuals are related to the observations through Equations (11.1), and the total probability of their simultaneous occurrence is given
by Equation (11.5). However, notice in Equation (11.2) that h2 ⫽ 1/22, and
since weights are inversely proportional to variances, they are directly proportional to h2. Thus, Equation (11.5) can be rewritten as
2
2
2
P ⫽ Kn(⌬v)ne⫺(w1v1⫹w2v2⫹䡠䡠䡠⫹wnvn)
(11.10)
To maximize P in Equation (11.10), the negative exponent must be minimized. To achieve this, the sum of the products of the weights times their
respective residuals squared must be minimized. This is the condition imposed
in weighted least squares adjustment. The condition of weighted least squares
adjustment in equation form is
w1v12 ⫹ w2v22 ⫹ 䡠 䡠 䡠 ⫹ wnvn2 ⫽ 兺 wv2 → minimum
(11.11)
Substituting the values for the residuals given in Equation (11.1) into Equation (11.11) yields
w1(M ⫺ z1)2 ⫹ w2(M ⫺ z2)2 ⫹ 䡠 䡠 䡠 ⫹ wn(M ⫺ zn)2 → minimum
(11.12)
The condition for a weighted least squares adjustment is: The most probable value for a quantity obtained from repeated observations having various
weights is that value which renders the sum of the weights times their respective squared residuals a minimum.
11.5
FUNCTIONAL MODEL
177
The minimum condition is imposed by differentiating Equation (11.12)
with respect to M and setting the resulting equation equal to zero. This yields
2w1(M ⫺ z1)(1) ⫹ 2w2(M ⫺ z2)(1) ⫹ 䡠 䡠 䡠 ⫹ 2wn(M ⫺ zn)(1) ⫽ 0
(11.13)
Dividing Equation (11.13) by 2 and rearranging results in
w1(M ⫺ z1) ⫹ w2(M ⫺ z2) ⫹ 䡠 䡠 䡠 ⫹ wn(M ⫺ zn) ⫽ 0
(11.14a)
Rearranging Equation (11.14a) gives
w1z1 ⫹ w2z2 ⫹ 䡠 䡠 䡠 ⫹ wnzn ⫽ w1M ⫹ w2M ⫹ 䡠 䡠 䡠 ⫹ wn M
(11.14b)
Equation (11.14b) can be written as 兺 wz ⫽ 兺 wM. Thus,
M⫽
兺 wz
兺w
(11.15)
Notice that Equation (11.15) is the same as Equation (10.13), which is the
formula for computing the weighted mean.
11.4
STOCHASTIC MODEL
The determination of variances, and subsequently the weights of the observations, is known as the stochastic model in a least squares adjustment. In
Section 11.3 the inclusion of weights in the adjustment was discussed. It is
crucial to the adjustment to select a proper stochastic (weighting) model since,
as was discussed in Section 10.1, the weight of an observation controls the
amount of correction it receives during the adjustment. However, development
of the stochastic model is important not only to weighted adjustments. When
doing an unweighted adjustment, all observations are assumed to be of equal
weight, and thus the stochastic model is created implicitly. The foundations
for selecting a proper stochastic model in surveying were established in Chapters 7 to 10. It will be shown in Chapter 21 that failure to select the stochastic
model properly will also affect one’s ability to isolate blunders in observation
sets.
11.5
FUNCTIONAL MODEL
A functional model in adjustment computations is an equation or set of equations that represents or defines an adjustment condition. It must be either
known or assumed. If the functional model represents the physical situation
178
PRINCIPLES OF LEAST SQUARES
adequately, the observational errors can be expected to conform to the normal
distribution curve. For example, a well-known functional model states that
the sum of angles in a triangle is 180⬚. This model is adequate if the survey
is limited to a small region. However, when the survey covers very large
areas, this model does not account for the systematic errors caused by Earth’s
curvature. In this case, the functional model is inadequate and needs to be
modified to include corrections for spherical excess. In traversing, the functional model of plane computations is suitable for smaller surveys, but if the
extent of the survey becomes too large, the model must again be changed to
account for the systematic errors caused by Earth’s curvature. This can be
accomplished by transforming the observations into a plane mapping system
such as the state plane coordinate system or by using geodetic observation
equations. Needless to say, if the model does not fit the physical situation, an
incorrect adjustment will result. In Chapter 23 we discuss a three-dimensional
geodetic and the systematic errors that must be taken into account in a threedimensional geodetic adjustment.
There are two basic forms for functional models: the conditional and parametric adjustments. In a conditional adjustment, geometric conditions are
enforced on the observations and their residuals. Examples of conditional
adjustment are: (1) the sum of the angles in a closed polygon is (n ⫺ 2)180⬚,
where n is the number of sides in the polygon; (2) the latitudes and departures
of a polygon traverse sum to zero; and (3) the sum of the angles in the horizon
equal 360⬚. A least squares adjustment example using condition equations is
given in Section 11.13.
When performing a parametric adjustment, observations are expressed in
terms of unknown parameters that were never observed directly. For example,
the well-known coordinate equations are used to model the angles, directions,
and distances observed in a horizontal plane survey. The adjustment yields
the most probable values for the coordinates (parameters), which in turn provide the most probable values for the adjusted observations.
The choice of the functional model will determine which quantities or
parameters are adjusted. A primary purpose of an adjustment is to ensure that
all observations are used to find the most probable values for the unknowns
in the model. In least squares adjustments, no matter if conditional or parametric, the geometric checks at the end of the adjustment are satisfied and
the same adjusted observations are obtained. In complicated networks, it is
often difficult and time consuming to write the equations to express all conditions that must be met for a conditional adjustment. Thus, this book will
focus on the parametric adjustment, which generally leads to larger systems
of equations but is straightforward in its development and solution and, as a
result, is well suited to computers.
The combination of stochastic and functional models results in a mathematical model for the adjustment. The stochastic and functional models must
both be correct if the adjustment is to yield the most probable values for the
unknown parameters. That is, it is just as important to use a correct stochastic
11.6
179
OBSERVATION EQUATIONS
model as it is to use a correct functional model. Improper weighting of observations will result in the unknown parameters being determined incorrectly.
11.6
OBSERVATION EQUATIONS
Equations that relate observed quantities to both observational residuals and
independent unknown parameters are called observation equations. One equation is written for each observation and for a unique set of unknowns. For a
unique solution of unknowns, the number of equations must equal the number
of unknowns. Usually, there are more observations (and hence equations) than
unknowns, and this permits determination of the most probable values for the
unknowns based on the principle of least squares.
11.6.1
Elementary Example of Observation Equation Adjustment
As an example of a least squares adjustment by the observation equation
method, consider the following three equations:
(1) x ⫹ y ⫽ 3.0
(2) 2x ⫺ y ⫽ 1.5
(11.16)
(3) x ⫺ y ⫽ 0.2
Equations (11.16) relate the two unknowns, x and y, to the quantities observed (the values on the right side of the equations). One equation is
redundant since the values for x and y can be obtained from any two of the
three equations. For example, if Equations (1) and (2) are solved, x would
equal 1.5 and y would equal 1.5, but if Equations (2) and (3) are solved, x
would equal 1.3 and y would equal 1.1, and if Equations (1) and (3) are
solved, x would equal 1.6 and y would equal 1.4. Based on the inconsistency
of these equations, the observations contain errors. Therefore, new expressions, called observation equations, can be rewritten that include residuals.
The resulting set of equations is
(4) x ⫹ y ⫺ 3.0 ⫽ v1
(5) 2x ⫺ y ⫺ 1.5 ⫽ v2
(11.17)
(6) x ⫺ y ⫺ 0.2 ⫽ v3
Equations (11.17) relate the unknown parameters to the observations and
their errors. Obviously, it is possible to select values of v1, v2, and v3 that
will yield the same values for x and y no matter which pair of equations are
used. For example, to obtain consistencies through all of the equations, ar-
180
PRINCIPLES OF LEAST SQUARES
bitrarily let v1 ⫽ 0, v2 ⫽ 0, and v3 ⫽ ⫺0.2. In this arbitrary solution, x would
equal 1.5 and y would equal 1.5, no matter which pair of equations is solved.
This is a consistent solution, however, there are other values for the v’s that
will produce a smaller sum of squares.
To find the least squares solution for x and y, the residual equations are
squared and these squared expressions are added to give a function, ƒ(x,y),
that equals 兺 v2. Doing this for Equations (11.17) yields
ƒ(x,y) ⫽
冘 v ⫽ (x ⫹ y ⫺ 3.0) ⫹ (2x ⫺ y ⫺ 1.5) ⫹ (x ⫺ y ⫺ 0.2)
2
2
2
2
(11.18)
As discussed previously, to minimize a function, its derivatives must be
set equal to zero. Thus, in Equation (11.18), the partial derivatives of Equation
(11.18) with respect to each unknown must be taken and set equal to zero.
This leads to the two equations
⭸ƒ(x,y)
⫽ 2(x ⫹ y ⫺ 3.0) ⫹ 2(2x ⫺ y ⫺ 1.5)( ) ⫹ 2(x ⫺ y ⫺ 0.2) ⫽ 0
⭸x
⭸ƒ(x,y)
⫽ 2(x ⫹ y ⫺ 3.0) ⫹ 2(2x ⫺ y ⫺ 1.5)(⫺1) ⫹ 2(x ⫺ y ⫺ 0.2)(⫺1) ⫽ 0
⭸y
(11.19)
Equations (11.19) are called normal equations. Simplifying them gives
reduced normal equations of
6x ⫺ 2y ⫺ 6.2 ⫽ 0
(11.20)
⫺2x ⫹ 3y ⫺ 1.3 ⫽ 0
Simultaneous solution of Equations (11.20) yields x equal to 1.514 and y
equal to 1.442. Substituting these adjusted values into Equations (11.17), numerical values for the three residuals can be computed. Table 11.1 provides
a comparison of the arbitrary solution to the least squares solution. The tab-
TABLE 11.1 Comparison of an Arbitrary and a Least Squares Solution
Arbitrary Solution
v1 ⫽ 0
v2 ⫽ 0
v3 ⫽ ⫺0.02
Least Squares Solution
v21 ⫽ 0.0
v22 ⫽ 0.0
v23 ⫽ 0.04
0.04
v1 ⫽ ⫺0.044
v2 ⫽ 0.085
v3 ⫽ ⫺0.128
v21 ⫽ 0.002
v22 ⫽ 0.007
v23 ⫽ 0.016
0.025
11.7
181
SYSTEMATIC FORMULATION OF THE NORMAL EQUATIONS
ulated summations of residuals squared shows that the least squares solution
yields the smaller total and thus the better solution. In fact, it is the most
probable solution for the unknowns based on the observations.
11.7
11.7.1
SYSTEMATIC FORMULATION OF THE NORMAL EQUATIONS
Equal-Weight Case
In large systems of observation equations, it is helpful to use systematic procedures to formulate the normal equations. In developing these procedures,
consider the following generalized system of linear observation equations having variables of (A, B, C, . . . , N):
a1A ⫹ b1B ⫹ c1C ⫹ 䡠 䡠 䡠 ⫹ n1N ⫽ l1 ⫹ v1
a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫽ l2 ⫹ v2
(11.21)
⯗
am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫽ lm ⫹ vm
The squares of the residuals for Equations (11.21) are
v21 ⫽ (a1A ⫹ b1B ⫹ c1C ⫹ 䡠 䡠 䡠 ⫹ n1N ⫺ l1)2
v22 ⫽ (a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫺ l2)2
(11.22)
⯗
v2m ⫽ (am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫺ lm)2
Summing Equations (11.22), the function ƒ(A,B,C, . . . , N) ⫽ 兺 v2 is obtained. This expresses the equal-weight least squares condition as
冘 v ⫽ (a A ⫹ b B ⫹ c C ⫹ 䡠 䡠 䡠 ⫹ n N ⫺ l )
2
2
1
1
1
1
1
⫹ (a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫺ l2)2
⫹ 䡠 䡠 䡠 ⫹ (am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫺ lm)2
(11.23)
According to least squares theory, the minimum for Equation (11.23) is
found by setting the partial derivatives of the function with respect to each
unknown equal to zero. This results in the normal equations
182
PRINCIPLES OF LEAST SQUARES
⭸
兺 v2
⭸A
⫽ 2(a1A ⫹ b1B ⫹ c1C ⫹ 䡠 䡠 䡠 ⫹ n1N ⫺ l1)a1
⫹ 2(a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫺ l2)a2 ⫹ 䡠 䡠 䡠
⫹ 2(am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫺ lm)am ⫽ 0
⭸
兺 v2
⭸B
⫽ 2(a1A ⫹ b1B ⫹ c1C ⫹ 䡠 䡠 䡠 ⫹ n1N ⫺ l1)b1
⫹ 2(a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫺ l2)b2 ⫹ 䡠 䡠 䡠
⫹ 2(am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫺ lm)bm ⫽ 0
⭸
兺 v2
⭸C
⫽ 2(a1A ⫹ b1B ⫹ c1C ⫹ 䡠 䡠 䡠 ⫹ n1N ⫺ l1)c1
(11.24)
⫹ 2(a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫺ l2)c2 ⫹ 䡠 䡠 䡠
⫹ 2(am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫺ lm)cm ⫽ 0
⯗
⭸
兺 v2
⭸N
⫽ 2(a1A ⫹ b1B ⫹ c1C ⫹ 䡠 䡠 䡠 ⫹ n1N ⫺ l1)n1
⫹ 2(a2A ⫹ b2B ⫹ c2C ⫹ 䡠 䡠 䡠 ⫹ n2N ⫺ l2)n2 ⫹ 䡠 䡠 䡠
⫹ 2(am A ⫹ bm B ⫹ cmC ⫹ 䡠 䡠 䡠 ⫹ nm N ⫺ lm)nm ⫽ 0
Dividing each expression by 2 and regrouping the remaining terms in
Equation (11.24) results in
(a21 ⫹ a22 ⫹ 䡠 䡠 䡠 ⫹ am2 )A ⫹ (a1b1 ⫹ a2b2 ⫹ 䡠 䡠 䡠 ⫹ ambm)B
⫹ (a1c1 ⫹ a2c2 ⫹ 䡠 䡠 䡠 ⫹ amcm)C ⫹ 䡠 䡠 䡠 ⫹ (a1n1 ⫹ a2n2 ⫹ 䡠 䡠 䡠 ⫹ amnm)N
⫺ (a1l1 ⫹ a2l2 ⫹ 䡠 䡠 䡠 ⫹ amlm) ⫽ 0
(b1a1 ⫹ b2a2 ⫹ 䡠 䡠 䡠 ⫹ bmam)A ⫹ (b12 ⫹ b22 ⫹ 䡠 䡠 䡠 ⫹ b2m)B
⫹ (b1c1 ⫹ b2c2 ⫹ 䡠 䡠 䡠 ⫹ bmcm)C ⫹ 䡠 䡠 䡠 ⫹ (b1n1 ⫹ b2n2 ⫹ 䡠 䡠 䡠 ⫹ bmnm)N
⫺ (b1l1 ⫹ b2l2 ⫹ 䡠 䡠 䡠 ⫹ bmlm) ⫽ 0
(11.25)
11.7
183
SYSTEMATIC FORMULATION OF THE NORMAL EQUATIONS
(c1a1 ⫹ c2a2 ⫹ 䡠 䡠 䡠 ⫹ cmam)A ⫹ (c1b1 ⫹ c2b2 ⫹ 䡠 䡠 䡠 ⫹ cmbm)B
2
⫹ (c21 ⫹ c22 ⫹ 䡠 䡠 䡠 ⫹ cm
)C ⫹ 䡠 䡠 䡠 ⫹ (c1n1 ⫹ c2n2 ⫹ 䡠 䡠 䡠 ⫹ cmnm)N
⫺ (c1l1 ⫹ c2l2 ⫹ 䡠 䡠 䡠 ⫹ cmlm) ⫽ 0
⯗
(n1a1 ⫹ n2a2 ⫹ 䡠 䡠 䡠 ⫹ nmam)A ⫹ (n1b1 ⫹ n2b2 ⫹ 䡠 䡠 䡠 ⫹ nmbm)B
⫹ (n1c1 ⫹ n2c2 ⫹ 䡠 䡠 䡠 ⫹ nmcm)C ⫹ 䡠 䡠 䡠 ⫹ (n21 ⫹ n22 ⫹ 䡠 䡠 䡠 ⫹ n2m)N
⫺ (n1l1 ⫹ n2l2 ⫹ 䡠 䡠 䡠 ⫹ nmlm) ⫽ 0
Generalized equations expressing normal Equations (11.25) are now written as
冉冘 冊 冉冘 冊 冉冘 冊
冉冘 冊 冉冘 冊 冉冘 冊
冉冘 冊 冉冘 冊 冉冘 冊
a2 A ⫹
ab B ⫹
ac C ⫹ 䡠 䡠 䡠 ⫹
ba A ⫹
b2 B ⫹
bc C ⫹ 䡠 䡠 䡠 ⫹
ca A ⫹
cb B ⫹
c2 C ⫹ 䡠 䡠 䡠 ⫹
冉冘 冊
冉冘 冊
冉冘 冊
冘 al
bn N ⫽ 冘 bl
cn N ⫽ 冘 cl
an N ⫽
(11.26)
⯗
冉冘 冊 冉冘 冊 冉冘 冊
na A ⫹
nb B ⫹
nc C ⫹ 䡠 䡠 䡠 ⫹
冉冘 冊
n2 N ⫽
冘 nl
In Equation (11.26) the a’s, b’s, c’s, . . . , n’s are the coefficients for the
unknowns A, B, C, . . . , N; the l values are the observations; and 兺 signifies
summation from i ⫽ 1 to m.
11.7.2
Weighted Case
In a manner similar to that of Section 11.7.1, it can be shown that normal
equations can be formed systematically for weighted observation equations
in the following manner:
冘 (wa )A ⫹ 冘 (wab)B ⫹ 冘 (wac)C ⫹ 䡠 䡠 䡠 ⫹ 冘 (wan)N ⫽ 冘 wal
冘 (wba)A ⫹ 冘 (wb )B ⫹ 冘 (wbc)C ⫹ 䡠 䡠 䡠 ⫹ 冘 (wbn)N ⫽ 冘 wbl
冘 (wca)A ⫹ 冘 (wcb)B ⫹ 冘 (wc )C ⫹ 䡠 䡠 䡠 ⫹ 冘 (wcn)N ⫽ 冘 wcl
2
2
2
⯗
冘 (wna)A ⫹ 冘 (wnb)B ⫹ 冘 (wnc)C ⫹ 䡠 䡠 䡠 ⫹ 冘 (wn )N ⫽ 冘 wnl
2
(11.27)
184
PRINCIPLES OF LEAST SQUARES
In Equation (11.27), w are the weights of the observations, l; the a’s, b’s, c’s,
. . . , n’s are the coefficients for the unknowns A, B, C, . . . , N; the l values
are the observations; and 兺 signifies summation from i ⫽ 1 to m.
Notice that the terms in Equations (11.27) are the same as those in Equations (11.26) except for the addition of the w’s which are the relative weights
of the observations. In fact, Equations (11.27) can be thought of as the general
set of equations for forming the normal equations, since if the weights are
equal, they can all be given a value of 1. In this case they will cancel out of
Equations (11.27) to produce the special case given by Equations (11.26).
11.7.3
Advantages of the Systematic Approach
Using the systematic methods just demonstrated, the normal equations can be
formed for a set of linear equations without writing the residual equations,
compiling their summation equation, or taking partial derivatives with respect
to the unknowns. Rather, for any set of linear equations, the normal equations
for the least squares solution can be written directly.
11.8
TABULAR FORMATION OF THE NORMAL EQUATIONS
Formulation of normal equations from observation equations can be simplified
further by handling Equations (11.26) and (11.27) in a table. In this way, a
large quantity of numbers can be manipulated easily. Tabular formulation of
the normal equations for the example in Section 11.4.1 is illustrated below.
First, Equations (11.17) are made compatible with the generalized form of
Equations (11.21). This is shown in Equations (11.28).
(7) x ⫹ y ⫽ 3.0 ⫹ v1
(8) 2x ⫺ y ⫽ 1.5 ⫹ v2
(11.28)
(9) x ⫺ y ⫽ 0.2 ⫹ v3
In Equations (11.28), there are two unknowns, x and y, with different coefficients for each equation. Placing the coefficients and the observations, l’s,
for each expression of Equation (11.28) into a table, the normal equations are
formed systematically. Table 11.2 shows the coefficients, appropriate products, and summations in accordance with Equations (11.26).
After substituting the appropriate values for 兺 a2, 兺 ab, 兺 b2, 兺 al, and
兺 bl from Table 11.2 into Equations (11.26), the normal equations are
6x ⫺ 2y ⫽ 6.2
⫺2x ⫹ 3y ⫽ 1.3
(11.29)
11.9
USING MATRICES TO FORM THE NORMAL EQUATIONS
185
TABLE 11.2 Tabular Formation of Normal Equations
Eq.
a
b
7
8
9
1
2
1
⫺1
⫺1
1
l
a2
ab
b2
al
bl
3.0
1.5
0.2
1
4
1
6
1
⫺2
⫺1
1
1
1
3
3.0
3.0
0.2
6.2
⫺1.5
⫺0.2
⫺2
3.0
1.3
Notice that Equations (11.29) are exactly the same as those obtained in Section 11.6 using the theoretical least squares method. That is, Equations (11.29)
match Equations (11.20).
11.9
USING MATRICES TO FORM THE NORMAL EQUATIONS
Note that the number of normal equations in a parametric least squares adjustment is always equal to the number of unknown variables. Often, the
system of normal equations becomes quite large. But even when dealing with
three unknowns, their solution by hand is time consuming. As a consequence,
computers and matrix methods as described in Appendixes A through C are
used almost always today. In the following subsections we present the matrix
methods used in performing a least squares adjustment.
11.9.1
Equal-Weight Case
To develop the matrix expressions for performing least squares adjustments,
an analogy will be made with the systematic procedures demonstrated in
Section 11.7. For this development, let a system of observation equations be
represented by the matrix notation
(11.30)
AX ⫽ L ⫹ V
where
A⫽
X⫽
冤冥
x1
x2
⯗
xn
冤
a11
a21
a12
a22
⯗
⯗
am1
am2
L⫽
䡠䡠䡠
䡠䡠䡠
⯗
䡠䡠䡠
冤冥
l1
l2
⯗
lm
a1n
a2n
⯗
冥
amn
V⫽
冤冥
v1
v2
⯗
vm
186
PRINCIPLES OF LEAST SQUARES
Note that the system of observation equations (11.30) is identical to Equations (11.21) except that the unknowns are x1, x2, . . . , xn instead of A, B,
. . . , N, and the coefficients of the unknowns are a11, a12, . . . , a1n instead
of a1, b1, . . . , n1.
Subjecting the foregoing matrices to the manipulations given in the following expression, Equation (11.31) produces the normal equations [i.e., matrix
Equation (11.31a) is exactly equivalent to Equations (11.26)]:
ATAX ⫽ ATL
(11.31a)
Equation (11.31a) can also be expressed as
NX ⫽ ATL
(11.31b)
The correspondence between Equations (11.31) and (11.26) becomes clear
if the matrices are multiplied and analyzed as follows:
ATA ⫽
冤
冥冤
a11 a21 䡠 䡠 䡠 am1
a12 a22 䡠 䡠 䡠 am2
⯗
⯗
a1n a2n
⯗
⯗
䡠 䡠 䡠 amn
a11 a12 䡠 䡠 䡠 a1n
a21 a22 䡠 䡠 䡠 a2n
⯗
⯗
am1 am2
⯗
⯗
䡠 䡠 䡠 amn
冥冤
⫽
n11 n12 䡠 䡠 䡠 n1n
n21 n22 䡠 䡠 䡠 a2n
⯗
⯗
⯗
⯗
䡠 䡠 䡠 ann
an1 an2
冥
⫽N
冘a l
冘a l
m
冤冥
i1 i
ATL ⫽
冤
冥冤 冥
a11 a21 䡠 䡠 䡠 am1
a12 a22 䡠 䡠 䡠 am2
⯗
⯗
a1n a2n
⯗
⯗
䡠 䡠 䡠 amn
l1
l2
i⫽1
m
⫽
⯗
lm
i2 i
i⫽1
⯗
冘a l
m
in i
i⫽1
The individual elements of the N matrix can be expressed in the following
summation forms:
冘a
⫽冘a a
m
2
i1
n11 ⫽
m
i⫽1
i2 i1
i⫽1
⯗
2
i2
in i1
i⫽1
m
䡠䡠䡠
冘a a
n2n
i2 in
i⫽1
䡠䡠䡠
⯗
䡠䡠䡠
nnn ⫽
m
nn2 ⫽
in i2
i⫽1
i1 in
i⫽1
i⫽1
⯗
冘a a
n1n ⫽
m
n22
m
nn1 ⫽
i1 i2
冘a a
⫽冘a a
m
䡠䡠䡠
i⫽1
m
n21
冘a a
⫽冘a
n12 ⫽
冘a
m
2
in
i⫽1
11.9
187
USING MATRICES TO FORM THE NORMAL EQUATIONS
By comparing the summations above with those obtained in Equations
(11.26), it should be clear that they are the same. Therefore, it is demonstrated
that Equations (11.31a) and (11.31b) produce the normal equations of a least
squares adjustment. By inspection, it can also be seen that the N matrix is
always symmetric (i.e., nij ⫽ nji).
By employing matrix algebra, the solution of normal equations such as
Equation (11.31a) is
X ⫽ (ATA)⫺1ATL ⫽ N⫺1ATL
(11.32)
Example 11.1 To demonstrate this procedure, the problem of Section 11.6
will be solved. Equations (11.28) can be expressed in matrix form as
冤 冥冋 册 冤 冥 冤 冥
1
1
AX ⫽ 2 ⫺1
1 ⫺1
3.0
v1
x
⫽ 1.5 ⫹ v2 ⫽ L ⫹ V
y
0.2
v3
(a)
Applying Equation (11.31) to Equation (a) yields
ATAX ⫽ NX ⫽
冋
ATL ⫽
冋
1
1
册冤 冥冋 册 冋
册冤 冥 冋 册
2
⫺1
1
⫺1
1
2
1
1 ⫺1 ⫺1
1
1
2 ⫺1
1 ⫺1
x
6
⫽
y
⫺2
⫺2
3.0
6.2
1.5 ⫽
1.3
0.2
册冋 册
3
x
y
(b)
(c)
Finally, the adjusted unknowns, the X matrix, are obtained using the matrix
methods of Equation (11.32). This yields
X ⫽ N⫺1ATL ⫽
冋
6
⫺2
册冋 册 冋 册
⫺2
3
⫺1
6.2
1.514
⫽
1.3
1.442
(d)
Notice that the normal equations and the solution in this method are the same
as those obtained in Section 11.6.
11.9.2
Weighted Case
A system of weighted linear observation equations can be expressed in matrix
notation as
WAX ⫽ WL ⫹ WV
(11.33)
Using the methods demonstrated in Section 11.9.1, it is possible to show that
the normal equations for this weighted system are
188
PRINCIPLES OF LEAST SQUARES
ATWAX ⫽ ATWL
(11.34a)
Equation (11.34a) can also be expressed as
NX ⫽ ATWL
(11.34b)
where N ⫽ ATWA.
Using matrix algebra, the least squares solution of these weighted normal
equations is
X ⫽ (ATWA)⫺1ATWL ⫽ N⫺1ATWL
(11.35)
In Equation (11.35), W is the weight matrix as defined in Chapter 10.
11.10
LEAST SQUARES SOLUTION OF NONLINEAR SYSTEMS
In Appendix C we discuss a method of solving a nonlinear system of equations using a Taylor series approximation. Following this procedure, the least
squares solution for a system of nonlinear equations can be found as follows:
Step 1: Write the first-order Taylor series approximation for each equation.
Step 2: Determine initial approximations for the unknowns in the equations
of step 1.
Step 3: Use matrix methods similar to those discussed in Section 11.9 to find
the least squares solution for the equations of step 1 (these are corrections
to the initial approximations).
Step 4: Apply the corrections to the initial approximations.
Step 5: Repeat steps 1 through 4 until the corrections become sufficiently
small.
A system of nonlinear equations that are linearized by a Taylor series
approximation can be written as
JX ⫽ K ⫹ V
(11.36)
where the Jacobian matrix J contains the coefficients of the linearized observation equations. The individual matrices in Equation (11.36) are
11.10
LEAST SQUARES SOLUTION OF NONLINEAR SYSTEMS
J⫽
X⫽
冤冥 冤
dx1
dx2
⯗
K⫽
⭸F1
⭸x1
⭸F1
⭸x2
䡠䡠䡠
⭸F1
⭸xn
⭸F2
⭸x1
⭸F2
⭸x2
䡠䡠䡠
⭸F2
⭸xn
⯗
⯗
⯗
⯗
⭸Fm
⭸x1
⭸Fm
⭸x2
䡠䡠䡠
⭸Fm
⭸xn
l1 ⫺ ƒ1(x1, x2, . . . , xn)
l2 ⫺ ƒ2(x1, x2, . . . , xn)
⯗
冥
lm ⫺ ƒm(x1, x2, . . . , xn)
dxn
V⫽
189
冤冥
v1
v2
⯗
vm
The vector of least squares corrections in the equally weighted system of
Equation (11.36) is given by
X ⫽ (J TJ)⫺1J TK ⫽ N⫺1J TK
(11.37)
Similarly, the system of weighted equations is
WJX ⫽ WK
(11.38)
X ⫽ (J TWJ)⫺1J TWK ⫽ N⫺1J TWK
(11.39)
and its solution is
where W is the weight matrix as defined in Chapter 10. Notice that the least
squares solution of a nonlinear system of equations is similar to the linear
case. In fact, the only difference is the use of the Jacobian matrix rather than
the coefficient matrix and the use of the K matrix rather than the observation
matrix, L. Many authors use the same nomenclature for both the linear and
nonlinear cases. In these cases, the differences in the two systems of equations
are stated implicitly.
Example 11.2 Find the least squares solution for the following system of
nonlinear equations:
F: x ⫹ y ⫺ 2y2 ⫽ ⫺4
G: x2 ⫹ y2
H:
3x2 ⫺ y2
⫽
8
⫽
7.7
(e)
190
PRINCIPLES OF LEAST SQUARES
SOLUTION
Step 1: Determine the elements of the J matrix by taking partial derivatives
of Equation (e) with respect to the unknowns x and y. Then write the firstorder Taylor series equations.
⭸F
⫽1
⭸x
⭸G
⫽ 2x
⭸x
⭸H
⫽ 6x
⭸x
⭸F
⫽ 1 ⫺ 4y
⭸y
⭸G
⫽ 2y
⭸y
⭸H
⫽ ⫺2y
⭸y
冤
1
1 ⫺ 4y0
JX ⫽ 2x0
2y0
6x0
⫺2y0
冥冋 册 冤
冥
⫺4 ⫺ F(x0, y0)
8 ⫺ G(x0, y0)
7.7 ⫺ H(x0, y0)
dx
⫽
dy
⫽K
(ƒ)
Step 2: Determine initial approximations for the solution of the equations.
Initial approximations can be derived by solving any two equations for x
and y. This was done in Section C.3 for the equations for F and G, and
their solution was x0 ⫽ 2 and y0 ⫽ 2. Using these values, the evaluation
of the equations yields
F(x0,y0) ⫽ ⫺4
G(x0,y0) ⫽ 8
H(x0,y0) ⫽ 8
(g)
Substituting Equations (g) into the K matrix of Equation (ƒ), the K matrix
becomes
K⫽
冤
冥 冤 冥
⫺4 ⫺ (⫺4)
8⫺8
7.7 ⫺ 8
⫽
0
0
⫺0.3
It should not be surprising that the first two rows of the K matrix are zero
since the initial approximations were determined using these two equations.
In successive iterations, these values will change and all terms will become
nonzero.
Step 3: Solve the system using Equation (11.37).
N ⫽ J TJ ⫽
冋
册冤
J TK ⫽
冋
冥 冋
册
1 ⫺7
161 ⫺39
4
4 ⫽
⫺39
81
12 ⫺4
1 4 12
⫺7 4 ⫺4
册冤 冥 冋 册
1 4 12
⫺7 4 ⫺1
0
⫺3.6
0
⫽
1.2
⫺0.3
(h)
11.11
LEAST SQUARES FIT OF POINTS TO A LINE OR CURVE
191
Substituting the matrices of Equation (h) into Equation (11.37), the solution
for the first iteration is1
X ⫽ N⫺1J TK ⫽
冋
册
⫺0.02125
0.00458
Step 4: Apply the corrections to the initial approximations for the first
iteration.
x0 ⫽ 2.00 ⫺ 0.02125 ⫽ 1.97875
y0 ⫽ 2.00 ⫹ 0.00458 ⫽ 2.00458
Step 5: Repeating steps 2 through 4 results in
冋
册冋
157.61806 ⫺38.75082
X⫽N J K⫽
⫺38.75082
81.40354
⫺1 T
⫺1
册 冋
⫺0.017225
⫺0.003307
⫽
册
⫺0.00011
⫺0.00001
x ⫽ 1.97875 ⫺ 0.00011 ⫽ 1.97864
y ⫽ 2.00458 ⫺ 0.00001 ⫽ 2.00457
Iterating a third time yields extremely small corrections, and thus the final
solution, rounded to the hundredths place, is x ⫽ 1.98 and y ⫽ 2.00. Notice
that N changed by a relatively small amount from the first iteration to the
second iteration. If the initial approximations are close to their final values,
this can be expected. Thus, when doing these computations by hand, it is
common to use the initial N for each iteration, making it only necessary to
recompute JTK between iterations. However, this procedure should be used
with caution since if the initial approximations are poor, it will result in an
incorrect solution. One should always perform complete computations when
doing the solution with the aid of a computer.
11.11
LEAST SQUARES FIT OF POINTS TO A LINE OR CURVE
Frequently in engineering work, it is desirable or necessary to fit a straight
line or curve to a set of points with known coordinates. In solving this type
1
Note that although the solution represents more significant figures than can be warranted by the
observations, it is important to carry more digits than are desired for the final solution. Failure to
carry enough digits can result in a system that will never converge; rather, it may bounce above
and below the solution, or it may take more iterations, due to these rounding errors. This mistake
has been made by many beginning students. The answer should be rounded only after solving
the problem.
192
PRINCIPLES OF LEAST SQUARES
of problem, it is first necessary to decide on the appropriate functional model
for the data. The decision as to whether to use a straight line, parabola, or
some other higher-order curve can generally be made after plotting the data
and studying their form or by checking the size of the residuals after the least
squares solution with the first line or curve selected.
11.11.1
Fitting Data to a Straight Line
Consider the data illustrated in Figure 11.2. The straight line shown in the
figure can be represented by the equation
y ⫽ mx ⫹ b
(11.40)
In Equation (11.40), x and y are the coordinates of a point, m is the slope of
a line, and b is the y intercept at x ⫽ 0. If the points were truly linear and
there were no observational or experimental errors, all coordinates would lie
on a straight line. However, this is rarely the case, as shown in Figure 11.2,
and thus it is possible that (1) the points contain errors, (2) the functional
model is a higher-order curve, or both. If a line is selected as the model for
the data, the equation of the best-fitting straight line is found by adding residuals to Equations (11.40). This accounts for the errors shown in the figure.
Equations for the four data points A, B, C, and D of Figure 11.2 are rewritten
as
yA ⫹ vyA ⫽ mxA ⫹ b
yB ⫹ vyB ⫽ mxB ⫹ b
(11.41)
yC ⫹ vyC ⫽ mxC ⫹ b
yD ⫹ vyD ⫽ mxD ⫹ b
Equations (11.41) contain two unknowns, m and b, with four observations.
Their matrix representation is
Figure 11.2 Points on a line.
11.11
LEAST SQUARES FIT OF POINTS TO A LINE OR CURVE
193
(11.42)
AX ⫽ L ⫹ V
where
冋册
冤 冥
1
1
1
1
xa
x
A⫽ b
xc
xd
冤冥
ya
y
L⫽ b
yc
yd
m
X⫽
b
V⫽
冤冥
vya
vyb
vyc
vyd
Equation (11.42) is solved by the least squares method using Equation
(11.32). If some data were more reliable than others, relative weights could
be introduced and a weighted least squares solution could be obtained using
Equation (11.35).
Example 11.3 Find the best-fit straight line for the following points, whose
x and y coordinates are given in parentheses.
A: (3.00, 4.50)
B: (4.25, 4.25)
C: (5.50, 5.50)
D: (8.00, 5.50)
SOLUTION Following Equations (11.41), the four observation equations
for the coordinate pairs are
3.00m ⫹ b ⫽ 4.50 ⫹ va
4.25m ⫹ b ⫽ 4.25 ⫹ vb
(i)
5.50m ⫹ b ⫽ 5.50 ⫹ vc
8.00m ⫹ b ⫽ 5.50 ⫹ vd
Rewriting Equations (i) into matrix form yields
冤 冥冋 册 冤 冥 冤 冥
3.00 1
4.25 1
5.50 1
8.00 1
4.50
vA
m
4.25
vB
⫽
⫹
b
5.50
vC
5.50
vD
( j)
To form the normal equations, premultiply matrices A and L of Equation ( j)
by AT and get
冋
121.3125
20.7500
册冋 册 冋
20.7500
4.0000
The solution of Equation (k) is
册
m
105.8125
⫽
b
19.7500
(k)
194
PRINCIPLES OF LEAST SQUARES
X⫽
冋册 冋
m
121.3125
⫽
b
20.7500
册冋
20.7500
4.0000
⫺1
册 冋 册
105.8125
0.246
⫽
19.7500
3.663
Thus, the most probable values for m and b to the nearest hundredth are 0.25
and 3.66, respectively. To obtain the residuals, Equation (11.30) is rearranged
and solved as
冤 冥冋
3.00 1
4.25 1
V ⫽ AX ⫺ L ⫽
5.50 1
8.00 1
11.11.2
册
冤 冥冤 冥
4.50
⫺0.10
0.246
4.25
0.46
⫺
⫽
3.663
5.50
⫺0.48
5.50
0.13
Fitting Data to a Parabola
For certain data sets or in special situations, a parabola will fit the situation
best. An example would be fitting a vertical curve to an existing roadbed.
The general equation of a parabola is
Ax2 ⫹ Bx ⫹ C ⫽ y
(11.43)
Again, since the data rarely fit the equation exactly, residuals are introduced. For the data shown in Figure 11.3, the following observation equations
can be written:
Ax2a ⫹ Bxa ⫹ C ⫽ ya ⫹ va
Ax2b ⫹ Bxb ⫹ C ⫽ yb ⫹ vb
Axc2 ⫹ Bxc ⫹ C ⫽ yc ⫹ vc
Ax2d ⫹ Bxd ⫹ C ⫽ yd ⫹ vd
Axe2 ⫹ Bxe ⫹ C ⫽ ye ⫹ ve
Figure 11.3 Points on a parabolic curve.
(11.44)
11.12
CALIBRATION OF AN EDM INSTRUMENT
195
Equations (11.44) contain three unknowns, A, B, and C, with five equations.
Thus, this represents a redundant system that can be solved using least
squares. In terms of the unknown coefficients, Equations (11.44) are linear
and can be represented in matrix form as
AX ⫽ L ⫹ V
(11.45)
Since this is a linear system, it is solved using Equation (11.32). If weights
were introduced, Equation (11.35) would be used. The steps taken would be
similar to those used in Section 11.11.1.
11.12
CALIBRATION OF AN EDM INSTRUMENT
Calibration of an EDM is necessary to ensure confidence in the distances it
measures. In calibrating these devices, if they internally make corrections and
reductions for atmospheric conditions, Earth curvature, and slope, it is first
necessary to determine if these corrections are made properly. Once these
corrections are applied properly, the instruments with their reflectors must be
checked to determine their constant and scaling corrections. This is often
accomplished using a calibration baseline. The observation equation for an
electronically observed distance on a calibration baseline is
SDA ⫹ C ⫽ DH ⫺ DA ⫹ VDH
(11.46)
In Equation (11.46), S is a scaling factor for the EDM; C is an instrument–
reflector constant; DH is the horizontal distance observed with all atmospheric
and slope corrections applied; DA is the published horizontal calibrated distance for the baseline; and VDH is the residual error for each observation. This
is a linear equation with two unknowns, S and C. Systems of these equations
can be solved using Equation (11.31).
Example 11.4 A surveyor wishes to use an instrument–reflector combination that has an unknown constant value. Calibration baseline observations
were made carefully, and following the manufacturer’s recommendations, the
necessary corrections were applied for the atmospheric conditions, Earth curvature, and slope. Use these corrected distances and their published values,
listed in Table 11.3, to determine the instrument–reflector constant (C) and
scaling factor (S) for the system.
196
PRINCIPLES OF LEAST SQUARES
TABLE 11.3 EDM Instrument–Reflector Calibration Data
Distance
DA (m)
DH (m)
Distance
DA (m)
DH (m)
0–150
0–430
0–1400
150–430
150–1400
430–1400
149.9975
430.0101
1400.003
280.0126
1250.0055
969.9929
150.0175
430.0302
1400.0223
280.0327
1250.0248
970.0119
150–0
430–0
1400–0
430–150
1400–150
430–1400
149.9975
430.0101
1400.003
280.0126
1250.0055
969.9929
150.0174
430.0304
1400.0221
280.0331
1250.0257
970.0125
SOLUTION Following Equation (11.46), the matrix equation for this problem is
149.9975
149.9975
430.0101
430.0101
1400.0030
1400.0030
280.0126
280.0126
1250.0055
1250.0055
969.9929
969.9929
1
150.0175 ⫺ 149.9975
1
150.0174 ⫺ 149.9975
1
430.0302 ⫺ 430.0101
1
430.0304 ⫺ 430.0101
1
1400.0223 ⫺ 1400.0030
1 S
1400.0221 ⫺ 1400.0030
⫽
⫹V
1 C
280.0327 ⫺ 280.0126
1
280.0331 ⫺ 280.0126
1
1250.0248 ⫺ 1250.0055
1
1250.0257 ⫺ 1250.0055
1
970.0119 ⫺ 969.9929
1
970.0125 ⫺ 969.9929
冋册
Using Equation (11.32), the solution is S ⫽ ⫺0.0000007 (⫺0.7 ppm) and
C ⫽ 0.0203. Thus, the constant value for the instrument–reflector pair is
approximately 0.020 m, or 20 mm.
11.13 LEAST SQUARES ADJUSTMENT USING
CONDITIONAL EQUATIONS
As stated in Section 11.5, observations can also be adjusted using conditional
equations. In this section this form of adjustment is demonstrated by using
the condition that the sum of the angles in the horizon at a single station must
equal 360⬚.
Example 11.5 While observing angles at a station, the horizon was closed.
The following observations and their standard deviations were obtained:
11.13
No.
a1
a2
a3
LEAST SQUARES ADJUSTMENT USING CONDITIONAL EQUATIONS
Angle
S (ⴖ)
134⬚38⬘56ⴖ
83⬚17⬘35ⴖ
142⬚03⬘14ⴖ
Ⳳ6.7
Ⳳ9.9
Ⳳ4.3
197
What are the most probable values for these observations?
SOLUTION In a conditional adjustment, the most probable set of residuals
are found that satisfy a given functional condition. In this case, the condition
is that the sum of the three angles is equal to 360⬚. Since the three angles
observed actually sum to 359⬚59⬘45ⴖ, the angular misclosure is 15ⴖ. Thus,
errors are present. The following residual equations are written for the observations listed above.
v1 ⫹ v2 ⫹ v3 ⫽ 360⬚ ⫺ (a1 ⫹ a2 ⫹ a3) ⫽ 15ⴖ
(l)
In Equation (l), the a’s represent the observations and the v’s are residuals.
Applying the fundamental condition for a weighted least squares adjustment, the following equation must be minimized:
F ⫽ w1v12 ⫹ w2v22 ⫹ w3v32
(m)
where the w’s are weights, which are the inverses of the squares of the
standard deviations.
Equation (l) can be rearranged such that v3 is expressed as a function of
the other two residuals, or
v3 ⫽ 15 ⫺ (v1 ⫹ v2)
(n)
Substituting Equation (n) into Equation (m) yields
F ⫽ w1v21 ⫹ w2v22 ⫹ w3[15 ⫺ (v1 ⫹ v2)]2
(o)
Taking the partial derivatives of F with respect to both v1 and v2, respectively,
in Equation (o) results in the following two equations:
⭸F
⫽ 2w1v1 ⫹ 2w3[15ⴖ ⫺ (v1 ⫹ v2)](⫺1) ⫽ 0
⭸v1
(p)
⭸F
⫽ 2w2v2 ⫹ 2w3[15ⴖ ⫺ (v1 ⫹ v2)](⫺1) ⫽ 0
⭸v2
Rearranging Equations (p) and substituting in the appropriate weights yields
the following normal equations:
198
PRINCIPLES OF LEAST SQUARES
冉
冊
冉 冊
冉 冊
1
1
1
1
⫹
v ⫹
v ⫽ 15
6.72 4.32 1 4.32 2
4.32
冉
冊
(q)
1
1
1
1
v ⫹
⫹
v ⫽ 15
4.32 1
9.92 4.32 2
4.32
Solving Equations (q) for v1 and v2 yields
v1 ⫽ 4.2ⴖ
v2 ⫽ 9.1ⴖ
By substituting these residual values into Equation (n), residual v3 is computed as
v3 ⫽ 15ⴖ ⫺ (4.2ⴖ ⫹ 9.1ⴖ) ⫽ 1.7ⴖ
Finally, the adjusted observations are obtained by adding to the observations
the residuals that were computed.
No.
a1
a2
a3
Observed Angle
v (ⴖ)
Adjusted Angle
134⬚38⬘56ⴖ
83⬚17⬘35ⴖ
142⬚03⬘14ⴖ
4.2
9.1
1.7
134⬚39⬘00.2ⴖ
83⬚17⬘44.1ⴖ
142⬚03⬘15.7ⴖ
360⬚00⬘00.0ⴖ
Note that geometric closure has been enforced in the adjusted angles to make
their sum exactly 360⬚. Note also that the angle having the smallest standard
deviation received the smallest correction (i.e., its residual is smallest).
11.14
EXAMPLE 11.5 USING OBSERVATION EQUATIONS
Example 11.5 can also be done using observation equations. In this case, the
three observations are related to their adjusted values and their residuals by
writing observation equations
11.14
EXAMPLE 11.5 USING OBSERVATION EQUATIONS
199
a1 ⫽ 134⬚38⬘56ⴖ ⫹ v1
a2 ⫽ 83⬚17⬘35ⴖ ⫹ v2
(r)
a3 ⫽ 142⬚03⬘14ⴖ ⫹ v3
While these equations relate the adjusted observations to their observed
values, they cannot be solved in this form. What is needed is the constraint,2
which states that the sum of the three angles equals 360⬚. This equation is
a1 ⫹ a2 ⫹ a3 ⫽ 360⬚
(s)
Rearranging Equation (s) to solve for a3 yields
a3 ⫽ 360⬚ ⫺ (a1 ⫹ a2)
(t)
Substituting Equation (t) into Equations (r) produces
a1 ⫽ 134⬚38⬘56ⴖ ⫹ v1
a2 ⫽ 83⬚17⬘35ⴖ ⫹ v2
(u)
360⬚ ⫺ (a1 ⫹ a2) ⫽ 142⬚03⬘14ⴖ ⫹ v3
This is a linear problem with two unknowns, a1 and a2. The weighted
observation equation solution is obtained by solving Equation (11.35). The
appropriate matrices for this problem are
冤
1
0
A⫽
0
1
⫺1 ⫺1
冥
W⫽
冤 冥
1
6.72
0
0
0
1
9.92
0
0
1
4.32
0
L⫽
冤
冥
134⬚38⬘56ⴖ
83⬚17⬘35ⴖ
142⬚03⬘14ⴖ ⫺ 360⬚
Performing matrix manipulations, the coefficients of the normal equations are
2
Chapter 20 covers the use of constraint equations in least squares adjustment.
200
PRINCIPLES OF LEAST SQUARES
ATWA ⫽
冋
冤 冥冤
册
1 0 ⫺1
0 1 ⫺1
1
6.72
0
0
0
1
9.92
0
0
1
4.32
0
⫽
ATWL ⫽
冋
冋
0.07636
0.050408
册
冥
1 0
0 1
⫺1 1
0.05408
0.06429
册
14.7867721
12.6370848
Finally, X is computed as
X ⫽ (ATWA)⫺1ATWL ⫽
冋
册
134⬚39⬘00.2ⴖ
83⬚17⬘44.1ⴖ
Using Equation (t), it can now be determined that a3 is 360⬚ ⫺ 134⬚39⬘00.2ⴖ
⫺ 83⬚17⬘44.1ⴖ ⫽ 142⬚03⬘15.7ⴖ. The same result is obtained as in Section
11.13. It is important to note that no matter what method of least squares
adjustment is used, if the procedures are performed properly, the same solution will always be obtained. This example involved constraint equation (t).
This topic is covered in more detail in Chapter 20.
PROBLEMS
11.1
Calculate the most probable values for A and B in the equations below
by the method of least squares. Consider the observations to be of
equal weight. (Use the tabular method to form normal equations.)
(a) 3A ⫹ 2B ⫽ 7.80 ⫹ v1
(b) 2A ⫺ 3B ⫽ 5.55 ⫹ v2
(c) 6A ⫺ 7B ⫽ 8.50 ⫹ v3
11.2
If observations (a), (b), and (c) in Problem 11.1 have weights of 6,
4, and 3, respectively, solve the equations for the most probable values
of A and B using weighted least squares. (Use the tabular method to
form normal equations.)
11.3
Repeat Problem 11.1 using matrices.
11.4
Repeat Problem 11.2 using matrices.
PROBLEMS
201
11.5
Solve the following nonlinear equations using the least squares
method.
(a) x2 ⫹ 3xy ⫺ y2 ⫽ 16.0
(b) 7x3 ⫺ 3y2 ⫽ 71.7
(c) 2x ⫺ 6xy ⫹ 3y2 ⫽ 3.2
11.6
The following coordinates of points on a line were computed for a
block. What are the slope and y intercept of the line? What is the
azimuth of the line?
11.7
11.8
11.9
Point
X (ft)
Y (ft)
1
2
3
4
1254.72
1362.50
1578.94
1843.68
2951.76
3205.13
3713.80
4335.92
What are the most probable values for the three angles observed to
close the horizon at station Red. The observed values and their standard deviations are:
Angle
Value
S (ⴖ)
1
2
3
123⬚32⬘56ⴖ
110⬚07⬘28ⴖ
126⬚19⬘44ⴖ
Ⳳ2.5
Ⳳ1.5
Ⳳ4.9
Determine the most probable values for the three interior of a triangle
that were measured as:
Angle
Value
S (ⴖ)
1
2
3
58⬚26⬘48ⴖ
67⬚06⬘56ⴖ
54⬚26⬘24ⴖ
Ⳳ5.1
Ⳳ4.3
Ⳳ2.6
Eight blocks of the Main Street are to be reconstructed. The existing
street consists of short, jogging segments as tabulated in the traverse
survey data below. Assuming coordinates of X ⫽ 1000.0 and Y ⫽
1000.0 at station A, and that the azimuth of AB is 90⬚, define a new
straight alignment for a reconstructed street passing through this area
202
PRINCIPLES OF LEAST SQUARES
which best conforms to the present alignment. Give the Y intercept
and the azimuth of the new alignment.
Course
Length (ft)
Station
Angle to Right
AB
BC
CD
DE
EF
FG
GH
HI
635.74
364.82
302.15
220.08
617.36
429.04
387.33
234.28
B
C
D
E
F
G
H
180⬚01⬘26ⴖ
179⬚59⬘52ⴖ
179⬚48⬘34ⴖ
180⬚01⬘28ⴖ
179⬚59⬘05ⴖ
180⬚01⬘37ⴖ
179⬚59⬘56ⴖ
11.10 Use the ADJUST software to do Problem 11.9.
11.11 The property corners on a single block with an alley are shown as a
straight line with a Due East bearing on a recorded plat. During a
recent survey, all the lot corners were found, and measurements from
station A to each were obtained. The surveyor wishes to determine
the possibility of disturbance of the corners by checking their fit to a
straight line. A sketch of the situation is shown in Figure P11.11, and
the results of the survey are given below. Assuming that station A has
coordinates of X ⫽ 5000.00 and Y ⫽ 5000.00 and that the coordinates
of the backsight station are X ⫽ 5000.10 and Y ⫽ 5200.00, determine
the best-fitting line for the corners. Give the Y intercept and the bearing of the best-fit line.
Course
Distance (ft)
Angle at A
AB
AC
AD
AE
AF
AG
AH
AI
100.02
200.12
300.08
399.96
419.94
519.99
620.04
720.08
90⬚00⬘16ⴖ
90⬚00⬘08ⴖ
89⬚59⬘48ⴖ
90⬚01⬘02ⴖ
89⬚59⬘48ⴖ
90⬚00⬘20ⴖ
89⬚59⬘36ⴖ
90⬚00⬘06ⴖ
Figure P11.11
PROBLEMS
203
11.12 Use the ADJUST software to do Problem 11.11.
11.13 Calculate a best-fit parabola for the following data obtained on a
survey of an existing vertical curve, and determine the deviation (residuals) of the road from this best-fit curve. The curve starts at station
10⫹00 and ends at station 18⫹00. List the adjusted station elevations
and their residuals.
Station
Elevation
Station
Elevation
10⫹00
11⫹00
12⫹00
13⫹00
14⫹00
51.2
49.5
48.2
47.3
46.8
15⫹00
16⫹00
17⫹00
18⫹00
46.9
47.3
48.3
49.6
11.14 Use the ADJUST software to do Problem 11.13.
11.15 Using a procedure similar to that in Section 11.7.1, derive Equations
(11.27).
11.16 Using a procedure similar to that used in Section 11.9.1, show that
the matrix operations in Equation (11.34) result in the normal equations for a linear set of weighted observation equations.
11.17 Discuss the importance of selecting the stochastic model when adjusting data.
11.18 The values for three angles in a triangle, observed using a total station
and the directional method, are
Angle
Number of
Repetitions
Value
A
B
C
2
3
6
14⬚25⬘20ⴖ
58⬚16⬘00ⴖ
107⬚19⬘10ⴖ
The observed lengths of the course are
AB ⫽ 971.25 ft
BC ⫽ 253.25 ft
CA ⫽ 865.28 ft
The following estimated errors are assumed for each measurement:
i ⫽ Ⳳ0.003 ft
t ⫽ Ⳳ0.020 ft
DIN ⫽ Ⳳ2.0ⴖ
What are the most probable values for the angles? Use the conditional
equation method.
204
PRINCIPLES OF LEAST SQUARES
11.19 Do Problem 11.18 using observation equations and a constraint as
presented in Section 11.13.
11.20 The following data were collected on a calibration baseline. Atmospheric refraction and Earth curvature corrections were made to the
measured distances, which are in units of meters. Determine the
instrument–reflector constant and any scaling factor.
Distance
DA
DH
Distance
DA
DH
0–150
0–430
0–1400
150–430
150–1400
430–1400
149.9104
430.001
1399.9313
280.0906
1250.0209
969.9303
149.9447
430.0334
1399.9777
280.1238
1250.0795
969.9546
150–0
430–0
1400–0
430–150
1400–150
1400–430
149.9104
430.001
1399.9313
280.0906
1250.0209
969.9303
149.9435
430.034
1399.9519
280.123
1250.0664
969.9630
11.21 A survey of the centerline of a horizontal metric curve is done to
determine the as-built curve specifications. The coordinates for the
points along the curve are:
Point
X (ft)
Y (ft)
1
2
3
4
5
6
10,006.82
10,013.12
10,024.01
10,032.44
10,038.26
10,041.39
10,007.31
10,015.07
10,031.83
10,049.95
10,069.04
10,088.83
(a) Using Equation (C.10), compute the most probable values for the
radius and center of the circle.
(b) If two points located on the tangents have coordinates of
(9987.36, 9987.40) and (10,044.09, 10,119.54), what are the coordinates of the PC and PT of the curve?
CHAPTER 12
ADJUSTMENT OF LEVEL NETS
12.1
INTRODUCTION
Differential leveling observations are used to determine differences in elevation between stations. As with all observations, these measurements are subject to random errors that can be adjusted using the method of least squares.
In this chapter the observation equation method for adjusting differential leveling observations by least squares is developed, and several examples are
given to illustrate the adjustment procedures.
12.2
OBSERVATION EQUATION
To apply the method of least squares in leveling adjustments, a prototype
observation equation is first written for any elevation difference. Figure 12.1
illustrates the functional relationship for the elevation difference observed
between two stations, I and J. The equation is expressed as
Ej ⫺ Ei ⫽ ⌬Elevij ⫹ v⌬Elevij
(12.1)
This prototype observation equation relates the unknown elevations of any
two stations, I and J, with the differential leveling observation ⌬Elevij and its
residual ⌬vElevij. This equation is fundamental in performing least squares
adjustments of differential level nets.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
205
206
ADJUSTMENT OF LEVEL NETS
Figure 12.1 Differential leveling observation.
12.3
UNWEIGHTED EXAMPLE
In Figure 12.2, a leveling network and its survey data are shown. Assume
that all observations are equal in weight. In this figure, arrows indicate the
direction of leveling, and thus for line 1, leveling proceeds from benchmark
X to A with an observed elevation difference of ⫹5.10 ft. By substituting into
prototype equation (12.1), an observation equation is written for each observation in Figure 12.2. The resulting equations are
BM X = 100.00
1
4
C
7
5
B
6
3
Line
A
1
2
3
4
5
6
7
2
BM Y = 107.50
Figure 12.2 Interlocking leveling network.
Observed
Elevation
Difference
5.10
2.34
–1.25
–6.13
–0.68
–3.00
1.70
12.3
UNWEIGHTED EXAMPLE
A
⫺BM X ⫽ 5.10 ⫹ v1
⫺A
⫹BM Y ⫽ 2.34 ⫹ v2
207
C ⫺BM Y ⫽ ⫺1.25 ⫹ v3
⫺C
⫺A
⫹BM X ⫽ ⫺6.13 ⫹ v4
B
⫽ ⫺0.68 ⫹ v5
B
⫺BM Y ⫽ ⫺3.00 ⫹ v6
⫺B
(12.2)
⫽ ⫹1.70 ⫹ v7
C
Rearranging so that the known benchmarks are on the right-hand side of
the equations and substituting in their appropriate elevations yields
A
⫽ ⫹105.10 ⫹ v1
⫺A
⫽ ⫺105.16 ⫹ v2
C ⫽ ⫹106.25 ⫹ v3
⫺C ⫽ ⫺106.13 ⫹ v4
⫺A
⫹B
B
⫺B
(12.3)
⫽ ⫺0.68 ⫹ v5
⫽ ⫹104.50 ⫹ v6
⫹C ⫽ ⫹1.70 ⫹ v7
In this example there are three unknowns, A, B, and C. In matrix form,
Equations (12.2) are written as
(12.4a)
AX ⫹ B ⫽ L ⫹ V
where
0
0
0
0
0
0
1
A⫽
0
0 ⫺1
⫺1
1
0
0
1
0
0 ⫺1
1
X⫽
A
B
C
⫺100.00
⫹107.50
⫺107.50
1
⫺1
B⫽
100.00
0
⫺107.50
0
208
ADJUSTMENT OF LEVEL NETS
5.10
2.34
⫺1.25
L ⫽ ⫺6.13
⫺0.68
⫺3.00
1.70
V⫽
v1
v2
v3
v4
v5
v6
v7
In Equation (12.4a), the B matrix is a vector of the constants (benchmarks)
collected from the left side of the equation and L is a collection of elevation
differences observed using differential leveling. The right side of Equation
(12.3) is equal to L ⫺ B. It is a collection of the constants in the observation
equations and is often referred to as the constants matrix, L where L is the
difference between the differential leveling observations and constants in B.
Since the benchmarks can also be thought of as observations, this combination
of benchmarks and differential leveling observations is referred to as L in this
book, and Equation (12.4a) is simplified as
(12.4b)
AX ⫽ L ⫹ V
Also note in the A matrix that when an unknown does not appear in an
equation, its coefficient is zero. Since this is an unweighted example, according to Equation (11.31) the normal equations are
3 1 0
ATA ⫽ NX ⫽ 1 3 1
0 1 3
A
B
C
210.94
and ATL ⫽ 102.12
214.08
(12.5)
Using Equation (11.32), the solution of Equation (12.5) is
3 ⫺1
X ⫽ N⫺1ATL ⫽ ⫺1
3
0 ⫺1
⫽
0
⫺1
3
⫺1
210.94
102.12
214.08
0.38095 0.14286 0.04762
0.14286 0.42857 0.14286
0.04762 0.14286 0.38095
210.94
105.14
102.12 ⫽ 104.48
214.08
106.19
(12.6)
From Equation (12.6), the most probable elevations for A, B, and C are
105.14, 104.48, and 106.19, respectively. The rearranged form of Equation
(12.4b) is used to compute the residuals as
V ⫽ AX ⫺ L
From Equation (12.7), the matrix solution for V is
(12.7)
12.4
12.4
209
1
⫺1
V⫽
WEIGHTED EXAMPLE
0
0
⫺1
0
0
0
0
0
0
0
1
0 ⫺1
1
0
1
0
⫺1
1
105.10
0.041
0.019
105.141
106.25
⫺0.062
104.483 ⫺ ⫺106.13 ⫽ ⫺0.058
106.188
⫺0.68
0.022
104.50
⫺0.017
1.70
0.005
⫺105.16
WEIGHTED EXAMPLE
In Section 10.6 it was shown that relative weights for adjusting level lines
are inversely proportional to the lengths of the lines:
w⫽
1
length
(12.8)
The application of weights to the level circuit’s least squares adjustment is
illustrated by including the variable line lengths for the unweighted example
of Section 12.3. These line lengths for the leveling network of Figure 12.2
and their corresponding relative weights are given in Table 12.1. For convenience, each length is divided into the constant 12, so that integer ‘‘relative
weights’’ were obtained. (Note that this is an unnecessary step in the adjustment.) The observation equations are now formed as in Section 12.3, except
that in the weighted case, each equation is multiplied by its weight.
w1( A
) ⫽ w1(⫹105.10) ⫹ w1v1
w2(⫺A
) ⫽ w2(⫺105.16) ⫹ w2v2
w3(
C) ⫽ w3(⫹106.25) ⫹ w3v3
w4(
⫺C) ⫽ w4(⫺106.13) ⫹ w4v4
w5(⫺A ⫹B
) ⫽ w5(⫺0.68) ⫹ w5v5
w6(
B
) ⫽ w6(⫹104.50) ⫹ w6v6
w7(
⫺B
(12.9)
⫹C) ⫽ w7(⫹1.70) ⫹ w7v7
After dropping the residual terms in Equation (12.9), they can be written in
terms of matrices as
210
ADJUSTMENT OF LEVEL NETS
TABLE 12.1 Weights for the Example in
Figure 12.2
Line
Length (miles)
1
2
3
4
5
6
7
4
3
2
3
2
2
2
12 / 4
12 / 3
12 / 2
12 / 3
12 / 2
12 / 2
12 / 2
0
4
0
0
0
0
0
0
0
6
0
0
0
0
0
0
0
4
0
0
0
0
0
0
0
6
0
0
⫽
⫽
⫽
⫽
⫽
⫽
⫽
3 0 0 0 0 0 0
0 4 0 0 0 0 0
0 0 6 0 0 0 0
0 0 0 4 0 0 0
0 0 0 0 6 0 0
0 0 0 0 0 6 0
0 0 0 0 0 0 6
3
0
0
⫽ 0
0
0
0
Relative Weights
1
0
0
0
0
0
1
0 ⫺1
1
0
1
0
⫺1
0
⫺1
0
0
⫺1
0
0
0
0
0
0
0
6
0
0
0
0
0
0
0
6
3
4
6
4
6
6
6
A
B
C
(12.10)
105.10
⫺105.16
106.25
⫺106.13
⫺0.68
104.50
1.70
Applying Equation (11.34), we find that the normal equations are
(ATWA)X ⫽ NX ⫽ ATWL
(12.11)
where
1
N⫽ 0
0
⫺1
0
0
1
⫺1
0
0
⫺1
0
0
0
0 ⫺1 0
0
0
0
1 1 ⫺1
1 ⫺1
0 0
1
3
0
0
0
0
0
0
0
4
0
0
0
0
0
0
0
0
0
0
1
13 ⫺6
0
0 ⫺1 ⫽ ⫺6 18 ⫺6
1
0
0 ⫺6 16
1
0
⫺1
1
0
0
6
0
0
0
0
0
0
0
4
0
0
0
0
0
0
0
6
0
0
0
0
0
0
0
6
0
0
0
0
0
0
0
6
12.5
ATWL ⫽
REFERENCE STANDARD DEVIATION
211
740.02
612.72
1072.22
By using Equation (11.35), the solution for the X matrix is
0.0933 0.0355 0.0133
X ⫽ N⫺1(ATWL) ⫽ 0.0355 0.0770 0.0289
0.0133 0.0289 0.0733
740.02
105.150
612.72 ⫽ 104.489
1072.22
106.197
(12.12)
Equation (12.7) is now used to compute the residuals as
1
0
0
0
0
0
0
1
V ⫽ AX ⫺ L ⫽
0
0 ⫺1
⫺1
1
0
0
1
0
0 ⫺1
1
⫺1
105.10
0.050
0.010
105.150
106.25
⫺0.053
104.489 ⫺ ⫺106.13 ⫽ ⫺0.067
106.197
⫺0.68
0.019
104.50
⫺0.011
1.70
0.008
⫺105.16
It should be noted that these adjusted values (X matrix) and residuals (V
matrix) differ slightly from those obtained in the unweighted adjustment of
Section 12.3. This illustrates the effect of weights on an adjustment. Although
the differences in this example are small, for precise level circuits it is both
logical and wise to use a weighted adjustment since a correct stochastic model
will place the errors back in the observations that probably produced the
errors.
12.5
REFERENCE STANDARD DEVIATION
Equation (10.20) expressed the standard deviation for a weighted set of observations as
S0 ⫽
wv2
n ⫺ 1
(12.13)
However, Equation (12.13) applies to a multiple set of observations for a
single quantity where each observation has a different weight. Often, observations are obtained that involve several unknown parameters that are related
functionally like those in Equations (12.3) or (12.9). For these types of observations, the standard deviation in the unweighted case is
212
ADJUSTMENT OF LEVEL NETS
冪m ⫺ n 冪
兺 v2
S0 ⫽
兺 v2
⫽
r
which in matrix form is S0 ⫽
冪
V TV
r
(12.14)
In Equation (12.14), 兺 v2 is expressed in matrix form as V TV, m is the number
of observations, and n is the number of unknowns. There are r ⫽ m ⫺ n
redundant measurements or degrees of freedom in the adjustment.
The standard deviation for the weighted case is
S0 ⫽
冪m ⫺ n 冪
兺 wv2
⫽
兺 wv2
r
which in matrix form is S0 ⫽
冪
V TWV
(12.15)
r
where 兺 wv2 in matrix form is V TWV.
Since these standard deviations relate to the overall adjustment and not a
single quantity, they are referred to as reference standard deviations. Computations of the reference standard deviations for both unweighted and
weighted examples are illustrated below.
12.5.1
Unweighted Example
In the example of Section 12.3, there are 7 ⫺ 3, or 4, degrees of freedom.
Using the residuals given in Equation (12.7) and using Equation (12.14), the
reference standard deviation in the unweighted example is
S0 ⫽
冪
(0.041)2 ⫹ (0.019)2 ⫹ (⫺0.062)2 ⫹ (⫺0.058)2 ⫹ (0.022)2
⫹ (⫺0.017)2 ⫹ (0.005)2
7⫺3
⫽ 0.05
(12.16)
This can be computed using the matrix expression in Equation (12.14) as
S0 ⫽
冪
V TV
r
⫽ [0.041
⫽
0.019 ⫺0.062 ⫺0.058
⫽ 0.05
0.010
4
0.022 ⫺0.017
0.041
0.019
⫺0.062
0.005] ⫺0.058
0.022
⫺0.017
0.005
(12.17)
12.6
12.5.2
213
ANOTHER WEIGHTED ADJUSTMENT
Weighted Example
Notice that the weights are used when computing the reference standard deviation in Equation (12.15). That is, each residual is squared and multiplied
by its weight, and thus the reference standard deviation computed using nonmatrix methods is
S0 ⫽
⫽
冪
3(0.050)2 ⫹ 4(0.010)2 ⫹ 6(⫺0.053)2 ⫹ 4(⫺0.067)2 ⫹ 6(0.019)2
⫹ 6(⫺0.011)2 ⫹ 6(0.008)2
7⫺3
⫽ 0.107
冪0.04598
4
(12.18)
It is left as an exercise to verify this result by solving the matrix expression
of Equation (12.15).
12.6
ANOTHER WEIGHTED ADJUSTMENT
Example 12.1 The level net shown in Figure 12.3 is observed with the
following results (the elevation differences and standard deviations are given
in meters, and the elevation of A is 437.596 m):
From
To
⌬Elev (m)
(m)
From
To
⌬Elev (m)
(m)
A
B
C
B
C
D
10.509
5.360
⫺8.523
0.006
0.004
0.005
D
B
A
A
D
C
⫺7.348
⫺3.167
0.003
0.004
0.012
15.881
What are the most probable values for the elevations of B, C, and D?
Figure 12.3 Differential leveling network for Example 12.1.
214
ADJUSTMENT OF LEVEL NETS
SOLUTION
Step 1: Write the observation equations without their weights:
(1) ⫹B
⫽ A ⫹ 10.509 ⫹ v1 ⫽ 448.105 ⫹ v1
(2) ⫺B ⫹C
⫽ 5.360 ⫹ v2
(3)
⫹D
⫽ ⫺8.523 ⫹ v3
(4)
⫺D
⫽ ⫺A ⫺ 7.348 ⫹ v4 ⫽ ⫺444.944 ⫹ v4
(5) ⫺B
⫹D
⫽ ⫺3.167 ⫹ v5
⫺C
(6)
⫽ A ⫽ 15.881 ⫹ v6 ⫽ 453.477 ⫹ v6
⫹C
Step 2: Rewrite observation equations in matrix form AX ⫽ L ⫹ V as
1
⫺1
0
0
⫺1
0
0
0
1
0
⫺1
1
0 ⫺1
0
1
1
0
448.105
v1
5.360
v2
A
⫺8.523
v3
B ⫽
⫹
⫺444.944
v4
C
⫺3.167
v5
453.477
v6
(12.19)
Step 3: In accordance with Equations (10.4) and (10.6), form the weight
matrix as
1
0.0062
0
0
0
0
0
0
1
0.0042
0
0
0
0
0
0
1
0.0052
0
0
0
0
0
0
1
0.0032
0
0
0
0
0
0
1
0.0042
0
0
0
0
0
0
1
0.0122
W⫽
(12.20)
12.6
ANOTHER WEIGHTED ADJUSTMENT
215
from which
冤
27,778
0
0
0
0
0
0 62,500
0
0
0
0
0
0 40,000
0
0
0
W⫽
0
0
0 111,111
0
0
0
0
0
0 62,500
0
0
0
0
0
0 6944
(12.21)
Step 4: Compute the normal equations using Equation (11.34):
(ATWA)X ⫽ NX ⫽ ATWL
(12.22)
where
152,778 ⫺62,500 ⫺62,500
N ⫽ ⫺62,500 109,444 ⫺40,000
⫺62,500 ⫺40,000 213,611
ATWL ⫽
X⫽
B
C
D
12,310,298.611
3,825,065.833
48,899,364.722
Step 5: Solving for the X matrix using Equation (11.35) yields
448.1087
X ⫽ 453.4685
444.9436
(12.23)
Step 6: Compute the residuals using the matrix expression V ⫽ AX ⫺ L:
448.1087
448.105
0.0037
5.3598
5.360
⫺0.0002
⫺8.5249
⫺8.523
⫺0.0019
V⫽
⫺
⫽
⫺444.9436
⫺444.944
0.0004
⫺3.1651
⫺3.167
0.0019
453.4685
453.477
⫺0.0085
(12.24)
Step 7: Calculate the reference standard deviation for the adjustment using
the matrix expression of Equation (12.15):
216
ADJUSTMENT OF LEVEL NETS
V TWV ⫽ [0.0037 ⫺0.0002 ⫺0.0019
0.0004
0.0019 ⫺0.0085]
冤
0.0037
W
⫺0.0002
⫺0.0019
0.0004
0.0019
⫺0.0085
⫽ [1.26976]
(12.25)
Since the number of system redundancies is the number of observations
minus the number of unknowns, r ⫽ 6 ⫺ 3 ⫽ 3, and thus
S0 ⫽
⫽ 0.6575
1.26976
3
(12.26)
Step 8: Tabulate the results showing both the adjusted elevation differences,
their residuals, and final adjusted elevations.
From
To
Adjusted
⌬Elev
Residual
Station
Adjusted
Elevation
A
B
C
D
B
A
B
C
D
A
D
C
10.513
5.360
⫺8.525
⫺7.348
⫺3.165
15.872
0.004
0.000
⫺0.002
0.000
0.002
⫺0.009
A
B
C
D
437.596
448.109
453.468
444.944
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
PROBLEMS
217
Figure P12.1
12.1
For the leveling network in Figure P12.1, calculate the most probable
elevations for X and Y. Use an unweighted least squares adjustment
with the observed values given in the accompanying table. Assume
units of feet.
Line
⌬Elev (ft)
1
2
3
4
5
⫹3.68
⫹2.06
⫹2.02
⫺2.37
⫺0.38
12.2
For Problem 12.1, compute the reference standard deviation and tabulate the adjusted observations and their residuals.
12.3
Repeat Problem 12.1 using ADJUST.
Figure P12.4
12.4
For the leveling network shown in Figure P12.4, calculate the most
probable elevations for X, Y, and Z. The observed values and line
218
ADJUSTMENT OF LEVEL NETS
lengths are given in the table. Apply appropriate weights in the
computations.
Line
Length (mi)
⌬Elev (ft)
1
2
3
4
5
6
3
3
1.5
1.5
1
2
⫹1.02
⫺0.95
⫹1.96
⫺1.99
⫺0.04
⫹0.05
12.5
For Problem 12.4, compute the reference standard deviation and tabulate the adjusted observations and their residuals.
12.6
Use ADJUST to solve Problem 12.4.
12.7
A line of differential level is run from benchmark Oak (elevation
753.01) to station 13⫹00 on a proposed alignment. It continued along
the alignment to 19⫹00. Rod readings were taken on stakes at each
full station. The circuit then closed on benchmark Bridge, which has
an elevation of 772.52 ft. The elevation differences observed are, in
order, ⫺3.03, 4.10, 4.03, 7.92, 7.99, ⫺6.00, ⫺6.02, and 2.98 ft. A
third tie between benchmark Rock (elevation of 772.39 ft) and station
16⫹00 is observed as ⫺6.34 ft. What are:
(a) the most probable values for the adjusted elevations?
(b) the reference standard deviation for the adjustment?
(c) the adjusted observations and their residuals?
12.8
Use ADJUST to solve Problem 12.7.
12.9
If the elevation of A is 257.891 m, adjust the following leveling data
using the weighted least squares method.
From
To
⌬Elev (m)
Distance (km)
A
B
C
D
E
F
G
A
C
F
B
C
D
E
F
G
A
C
F
D
5.666
48.025
3.021
⫺13.987
20.677
⫺32.376
⫺30.973
53.700
9.634
⫺6.631
1
4.5
6
2.5
5
7.6
2.4
5.8
4.3
3.8
PROBLEMS
219
(a) What are the most probable values for the elevations of the
stations?
(b) What is the reference standard deviation?
(c) Tabulate the adjusted observations and their residuals.
12.10 Use ADJUST to solve Problem 12.9.
12.11 If the elevation of station 1 is 2395.67 ft, use weighted least squares
to adjust the following leveling.
From
To
⌬Elev
(ft)
Distance
(mi)
1
3
5
1
2
6
8
8
2
4
6
7
7
3
5
4
37.17
34.24
⫺23.12
16.99
⫺19.99
⫺0.04
⫺51.96
⫺41.14
3.00
1.56
0.83
1.61
2.91
3.06
2.98
1.08
⌬Elev
From
To
(ft)
2
4
6
7
7
3
6
3
5
1
3
6
8
8
⫺9.20
⫺10.92
⫺28.06
11.21
10.89
74.93
74.89
Distance
(mi)
3.63
1.98
0.93
1.21
1.41
1.77
8.03
(a) What are the most probable values for the elevations for the
stations?
(b) What is the adjustment’s reference standard deviation?
(c) Tabulate the adjusted observations and their residuals.
12.12 Use ADJUST to solve Problem 12.11.
12.13 Precise procedures were applied with a level that can be read to within
0.4 mm/1 m. The line of sight was held to within 3ⴖ of horizontal, and the sight distances were approximately 50 m in length. Use
these specifications and Equation (9.20) to compute standard deviations and hence weights. The elevation of A is 100.000 m. Adjust the
network by weighted least squares.
⌬Elev
From
To
(m)
A
B
C
D
B
C
D
A
⫺16.672
⫺7.903
12.383
12.190
Number
of Setups
From
To
⌬Elev
(m)
16
25
37
26
M
C
M
A
D
M
B
M
⫺38.238
30.338
⫺13.676
26.058
Number
of Setups
23
16
38
19
(a) What are the most probable values for the elevations of the
stations?
220
ADJUSTMENT OF LEVEL NETS
(b) What is the reference standard deviation for the adjustment?
(c) Tabulate the adjusted observations and their residuals.
12.14 Repeat Problem 12.13 using the number of setups for weighting following the procedures discussed in Section 10.6.
12.15 In Problem 12.13, the estimated error in reading the rod is 1.4 mm/
km. The bubble sensitivity of the instrument is 12 mm/km and the
average sight distances are 50 m. What are:
(a) the estimated standard errors for the observations?
(b) the most probable values for the elevations of the stations?
(c) the reference variance for the adjustment?
(d) Tabulate the adjusted observations and their residuals.
12.16 Demonstrate that 兺 v2 ⫽ V TV.
12.17 Demonstrate that 兺 wv2 ⫽ V TWV.
Programming Problems
12.18 Write a program that reads a file of differential leveling observations
and writes the matrices A, W, and L in a format suitable for the
MATRIX program. Using this package, solve Problem 12.11.
12.19 Write a computational package that reads the matrices A, W, and L,
computes the least squares solution for the unknown station elevations, and writes a file of adjusted elevation differences and their
residuals. Using this package, solve Problem 12.11.
12.20 Write a computational package that reads a file of differential leveling
observations, computes the least squares solution for the adjusted station elevations, and writes a file of adjusted elevation differences and
their residuals. Using this package, solve Problem 12.11.
CHAPTER 13
PRECISION OF INDIRECTLY
DETERMINED QUANTITIES
13.1
INTRODUCTION
Following an adjustment, it is important to know the estimated errors in both
the adjusted observations and the derived quantities. For example, after adjusting a level net as described in Chapter 12, the uncertainties in both adjusted elevation differences and computed benchmark elevations should be
determined. In Chapter 5, error propagation formulas were developed for
indirectly measured quantities which were functionally related to observed
values. In this chapter, error propagation formulas are developed for the quantities computed in a least squares solution.
13.2
DEVELOPMENT OF THE COVARIANCE MATRIX
Consider an adjustment involving weighted observation equations like those
in the level circuit example of Section 12.4. The matrix form for the system
of weighted observation equation is
WAX ⫽ WL ⫹ WV
(13.1)
and the least squares solution of the weighted observation equations is given
by
X ⫽ (ATWA)⫺1ATWL
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
(13.2)
221
222
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
In this equation, X contains the most probable values for the unknowns,
whereas the true values are Xtrue. The true values differ from X by some small
amount ⌬X, such that
X ⫹ ⌬X ⫽ Xtrue
(13.3)
where ⌬X represents the errors in the adjusted values.
Consider now a small incremental change, ⌬L, in the measured values, L,
which changes X to its true value, X ⫹ ⌬X. Then Equation (13.2) becomes
X ⫹ ⌬X ⫽ (ATWA)⫺1ATW(L ⫹ ⌬L)
(13.4)
Expanding Equation (13.4) yields
X ⫹ ⌬X ⫽ (ATWA)⫺1ATWL ⫹ (ATWA)⫺1ATW ⌬L
(13.5)
Note in Equation (13.2) that X ⫽ (ATWA)⫺1ATWL, and thus subtracting this
from Equation (13.5) yields
⌬X ⫽ (ATWA)⫺1ATW ⌬L
(13.6)
Recognizing ⌬L as the errors in the observations, Equation (13.6) can be
rewritten as
⌬X ⫽ (ATWA)⫺1ATWV
(13.7)
B ⫽ (ATWA)⫺1ATW
(13.8)
⌬X ⫽ BV
(13.9)
Now let
then
Multiplying both sides of Equation (13.9) by their transposes results in
⌬X ⌬X T ⫽ (BV)(BV)T
(13.10)
Applying the matrix property (BV)T ⫽ V TBT to Equation (13.10) gives
⌬X ⌬X T ⫽ BVV TBT
The expanded left side of Equation (13.11) is
(13.11)
13.2
⌬X ⌬X T ⫽
⌬x21
⌬x2 ⌬x1
⌬x3 ⌬x1
⯗
⌬xn ⌬x1
DEVELOPMENT OF THE COVARIANCE MATRIX
⌬x1 ⌬x2
⌬x22
⌬x3 ⌬x2
⯗
⌬xn ⌬x2
⌬x1 ⌬x3
⌬x2 ⌬x3
⌬x23
⯗
⌬xn ⌬x3
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
⌬x1 ⌬xn
⌬x2 ⌬xn
⌬x3 ⌬xn
⯗
⌬xn2
䡠
䡠
䡠
䡠
䡠
223
(13.12)
Also, the expanded right side of Equation (13.11) is
v21
v2v1
B v3v1
⯗
vmv1
v1v2
v22
v3v2
⯗
vmv2
v1v3
v2v3
v32
⯗
vmv3
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
v1vm
v2vm
v3vm BT
⯗
2
vm
䡠
䡠
䡠
䡠
䡠
(13.13)
Assume that it is possible to repeat the entire sequence of observations
many times, say a times, and that each time a slightly different solution
occurs, yielding a different set of X’s. Averaging these sets, the left side of
Equation (13.11) becomes
冘 ⌬x 冘 ⌬x ⌬x
a
a
冘 ⌬x ⌬x 冘 ⌬x
2
1
1
a
冘(⌬X)(⌬X)
T
1
2
⫽
2
2
2
1
a
a
冘 ⌬x ⌬x 冘 ⌬x ⌬x
⯗
⯗
1
n
2
n
a
a
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
䡠䡠䡠
冘 ⌬x ⌬x
a
冘 ⌬x ⌬x
1
n
2
n
a
冘
(13.14)
⯗
⌬xn2
a
If a is large, the terms in Equation (13.14) are the variances and covariances as defined in Equation (6.7) and Equation (13.14) can be rewritten as
S 2x1 Sx1x2 䡠 䡠 䡠 Sx1xn
Sx2x1 S 2x2 䡠 䡠 䡠 Sx2xn
⯗
⯗
Sxnx1 Sxnx2
䡠䡠䡠
䡠䡠䡠
⯗
S 2xn
⫽ S 2x
Also, considering a sets of observations, Equation (13.13) becomes
(13.15)
224
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
冘v 冘vv
a
a
冘vv 冘v
B
2
1
1 2
2 1
2
2
a
冘v v 冘v v
a
2 m
a
䡠䡠䡠
⯗
m 1
1 m
䡠䡠䡠
a
⯗
冘vv
a
冘vv
䡠䡠䡠
m 2
冘v
(13.16)
2
m
䡠䡠䡠
a
BT
a
Recognizing the diagonal terms as variances of the quantities observed, s2l ,
off-diagonal terms as the covariances, S 2lilj, and the fact that the matrix is
symmetric, Equation (13.16) can be rewritten as
B
S 2l1
Sl2l1
Sl1l2
S l22
⯗
⯗
Slml1
Slml2
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
Sl1lm
Sl2lm
⯗
S
2
lm
BT
(13.17)
In Section 10.1 it was shown that the weight of an observation is inversely
proportional to its variance. Also, from Equation (10.5), the variance of an
observation of weight w can be expressed in terms of the reference variance
as
S2i ⫽
S 20
wi
(13.18)
Recall from Equation (10.3) that W ⫽ Q⫺1 ⫽ 20⌺⫺1. Therefore, ⌺ ⫽ 20W⫺1,
and substituting Equation (13.18) into matrix (13.17) and replacing 0 with
S0 yields
T
S 20 BW⫺1
ll B
(13.19)
Substituting Equation (13.8) into Equation (13.19) gives
S 20 BW⫺1BT ⫽ S 02(ATWA)⫺1ATWW⫺1W TA [(ATWA)⫺1]T
(13.20)
Since the matrix of the normal equations is symmetric, it follows that
[(ATWA)⫺1]T ⫽ (ATWA)⫺1
(13.21)
Also, since the weight matrix W is symmetric, W T ⫽ W, and thus Equation
(13.20) reduces to
13.3
NUMERICAL EXAMPLES
S 20(ATWA)⫺1(ATWA)(ATWA)⫺1 ⫽ S 02(ATWA)⫺1
225
(13.22)
Equation (13.15) is the left side of Equation (13.11), for which Equation
(13.22) is the right. That is,
S 2x ⫽ S 20(ATWA)⫺1 ⫽ S 02N⫺1 ⫽ S 20Qxx
(13.23)
In least squares adjustment, the matrix Qxx of Equation (13.23) is known
as the variance–covariance matrix, or simply the covariance matrix. Diagonal
elements of the matrix when multiplied by S 20 give variances of the adjusted
quantities, and the off-diagonal elements multiplied by S 20 yield covariances.
From Equation (13.23), the estimated standard deviation Si for any unknown
parameter computed from a system of observation equations is expressed as
Si ⫽ S0qxixi
(13.24)
where qxixi is the diagonal element (from the ith row and ith column) of the
Qxx matrix, which as noted in Equation (13.23), is equal to the inverse of the
matrix of normal equations. Since the normal equation matrix is symmetric,
its inverse is also symmetric, and thus the covariance matrix is a symmetric
matrix (i.e., element ij ⫽ element ji).1
13.3
NUMERICAL EXAMPLES
The results of the level net adjustment in Section 12.3 will be used to illustrate
the computation of estimated errors for the adjusted unknowns. From Equation (12.6), the N⫺1 matrix, which is also the Qxx matrix, is
0.38095 0.14286 0.04762
Qxx ⫽ 0.14286 0.42857 0.14286
0.04762 0.14286 0.38095
Also, from Equation (12.17), S0 ⫽ Ⳳ0.05. Now by Equation (13.24), the
estimated standard deviations for the unknown benchmark elevations A, B,
and C are
1
Note that an estimate of the reference variance, 20, may be computed using either Equation
(12.13) or (12.14). However, it should be remembered that this only gives an estimate of the a
priori (before the adjustment) value for the reference variance. The validity of this estimate can
be checked using a 2 test as discussed in Chapter 5. If it is a valid estimate for 20, the a priori
value for the reference variance should be used in the computations discussed in this and subsequent chapters. Thus, if the a priori value for 20 is known, it should be used when computing
the a posteriori (after the adjustment) statistics. When weights are determined as 1 / 2i , the implicit
assumption made is that the a priori value for 20 ⫽ 1 [see Equations (10.5) and (10.6)].
226
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
SA ⫽ S0qAA ⫽ Ⳳ0.050.38095 ⫽ Ⳳ0.031 ft
SB ⫽ S0qBB ⫽ Ⳳ0.050.42857 ⫽ Ⳳ0.033 ft
SC ⫽ S0qCC ⫽ Ⳳ0.050.38095 ⫽ Ⳳ0.031 ft
In the weighted example of Section 12.4, it should be noted that although
this is a weighted adjustment, the a priori value for the reference variance is
not known because weights were determined as 1/distance and not 1/ 2i .
From Equation (12.12), the Qxx matrix is
0.0933 0.0355 0.0133
QXX ⫽ 0.0355 0.0770 0.0289
0.0133 0.0289 0.0733
Recalling that in Equation (12.18), S0 ⫽ Ⳳ0.107, the estimated errors in the
computed elevations of benchmarks A, B, and C are
SA ⫽ S0qAA ⫽ Ⳳ0.070.0933 ⫽ Ⳳ0.033 ft
SB ⫽ S0qBB ⫽ Ⳳ0.070.0770 ⫽ Ⳳ0.030 ft
SC ⫽ S0qCC ⫽ Ⳳ0.070.0733 ⫽ Ⳳ0.029 ft
These standard deviations are at the 68% probability level, and if other percentage errors are desired, these values should be multiplied by their respective t values as discussed in Chapter 3.
13.4
STANDARD DEVIATIONS OF COMPUTED QUANTITIES
In Section 6.1 the generalized law of propagation of variances was developed.
Recalled here for convenience, Equation (6.13) was written as
⌺ˆˆll ⫽ A⌺xxAT
where l̂ represents the adjusted observations, ⌺ˆˆll the covariance matrix of the
adjusted observations, ⌺xx the covariance matrix of the unknown parameters
[i.e., (ATWA)⫺1], and A, the coefficient matrix. Rearranging Equation (10.2)
and using sample statistics, there results ⌺xx ⫽ S 20Qxx. Also, from Equation
(13.23), S 2x ⫽ S 20Qxx ⫽ S 20(ATWA)⫺1 and thus ⌺xx ⫽ S 2x . Substituting this equality into Equation (a), the estimated standard deviations of the adjusted observations is
13.4
STANDARD DEVIATIONS OF COMPUTED QUANTITIES
⌺2ˆˆll ⫽ S 2ˆl ⫽ A⌺xx AT ⫽ AS 20(ATWA)⫺1AT ⫽ S 20AQxx AT ⫽ S 02Qˆˆll
227
(13.25)
where AQxx AT ⫽ Qˆˆll, which is known as the covariance matrix of the adjusted
observations.
Example 13.1 Consider the linear example in Section 12.3. By Equation
(13.25), the estimated standard deviations in the adjusted elevation differences
are
1
⫺1
S 2l̂ ⫽ 0.0502
⫻
1
0
0
0
0
⫺1
0
0
⫺1
0
0
0
0
0
0
0
1
0 ⫺1
1
0
1
0
⫺1
1
0
0
0
0
1 ⫺1
0.38095 0.14286 0.04762
0.14286 0.42857 0.14286
0.04762 0.14286 0.38095
⫺1
1
0
0
0
1 ⫺1
0
1
(13.26)
Performing the required matrix multiplications in Equation (13.26) yields
S 2l̂ ⫽ 0.0502
0.38095 ⫺0.38095
0.04762
0.38095 ⫺0.04762
0.04762 ⫺0.04762
0.38095
⫻ ⫺0.04762
0.04762 ⫺0.38095
⫺0.23810
0.23810
0.09524
0.14286 ⫺0.14286
0.14286
⫺0.09524
0.09524
0.23810
⫺0.38095
⫺0.04762
⫺0.23810
0.14286
0.04762
0.23810 ⫺0.14286
⫺0.38095
0.09524
0.14286
0.38095 ⫺0.09524 ⫺0.14286
⫺0.09524
0.52381
0.28571
⫺0.14286
0.28571
0.42857
⫺0.23810 ⫺0.19048 ⫺0.28571
⫺0.09524
0.09524
0.23810
⫺0.23810
⫺0.19048
⫺0.28571
0.52381
(13.27)
The estimated standard deviation of an observation is found by taking the
square root of the corresponding diagonal element of the S 2l̂ matrix (leveling
from A to B). For instance, for the fifth observation, Sl̂(5,5) applies and the
estimated error in the adjusted elevation difference of that observation is
S⌬ AB ⫽ Ⳳ0.0500.52381 ⫽ Ⳳ0.036 ft
An interpretation of the meaning of the value just calculated is that there
is a 68% probability that the true value is within the range Ⳳ0.036 ft of the
adjusted elevation difference (l5 ⫹ v5 ⫽ ⫺0.68 ⫹ 0.022 ⫽ ⫺0.658). That
228
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
is, the true value lies between ⫺0.694 and ⫺0.622 ft with 68% probability.
Careful examination of the matrix manipulations involved in solving Equation (13.25) for Example 13.1 reveals that the effort can be reduced significantly. In fact, to obtain the estimated standard deviation in the fifth element,
only the fifth row of the coefficient matrix, A, which represents the elevation
difference between A and B, need be used in the calculations. That row is
[⫺1 1 0]. Thus, to compute the standard deviation in this observation, the
following computations could be made:
S 2⌬ AB ⫽ 0.0502[⫺1
0.38095 0.14286 0.04762
0] 0.14286 0.42857 0.14286
0.04762 0.14286 0.38095
1
⫺1
1
0
⫺1
⫽ 0.0502[0.2389
⫺0.28571
⫺0.09524]
1
0
(13.28)
2
⫽ 0.050 [0.52381]
S⌬ AB ⫽ Ⳳ0.0500.52381 ⫽ Ⳳ0.036 ft
Note that this shortcut method produces the same value. Furthermore, because of the zero in the third position of this row from the coefficient matrix,
the matrix operations in Equation (13.28) could be further reduced to
S 2⌬ AB ⫽ 0.0502[⫺1
1]
0.38095 0.14286
0.14286 0.42857
⫺1
1
⫽ 0.0502[0.52381]
Another use for Equation (13.25) is in the computation of adjusted uncertainties for observations that were never made. For instance, in the example
of Section 12.3, the elevation difference between benchmarks X and B was
not observed. But from the results of the adjustment, this elevation difference
is 104.48 ⫺ 100.00 ⫽ 4.48 ft. The estimated error in this difference can be
found by writing an observation equation for it (i.e., B ⫽ X ⫹ ⌬ElevXB). This
equation does not involve either A or C, and thus in matrix form this difference would be expressed as
[0
1
0]
(13.29)
Using this row matrix in the same procedure as in Equation (13.28) yields
PROBLEMS
S 2⌬ XB ⫽ 0.0502[0
1
229
0.38095 0.14286 0.04762
0] 0.14286 0.42857 0.14286
0.04762 0.14286 0.38095
0
1
0
⫽ 0.0502[0.42857]
Hence,
S⌬ XB ⫽ Ⳳ0.0500.42857 ⫽ Ⳳ0.033 ft
Again, recognizing the presence of the zeros in the row matrix, these computations can be simplified to
S 2⌬ XB ⫽ 0.0501[1][0.42857][1] ⫽ 0.0502[0.42857]
The method illustrated above of eliminating unnecessary matrix computations
is formally known as matrix partitioning.
Computing uncertainties of quantities that were not actually observed has
application in many areas. For example, suppose that in a triangulation adjustment, the x and y coordinates of stations A and B are calculated and the
covariance matrix exists. Equation (13.25) could be applied to determine the
estimated error in the length of line AB calculated from the adjusted coordinates of A and B. This is accomplished by relating the length AB to the
unknown parameters as
AB ⫽ (Xb ⫺ Xa)2 ⫹ (Yb ⫺ Ya)2
(13.30)
This subject is discussed further in Chapter 15.
An important observation that should be made about the Qˆˆll and Qxx matrices is that only the coefficient matrix, A, is used in their formation. Since
the A matrix contains coefficients that express the relationships of the unknowns to each other, it depends only on the geometry of the problem. The
only other term in Equation (13.25) is the reference variance, and that depends
on the quality of the measurements. These are important concepts that will
be revisited in Chapter 21 when simulation of surveying networks is discussed.
PROBLEMS
For each problem, calculate the estimated errors for the adjusted benchmark
elevations.
230
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
13.1
The reference variance of an adjustment is 0.89. The covariance matrix and unknown parameter matrix are
0.5486
0.1864
0.0937
Qxx ⫽ 0.1864
0.4987 ⫺0.1678
0.0937 ⫺0.1678
0.8439
X⫽
A
B
C
What is the estimated error in the adjusted value for:
(a) A?
(b) B?
(c) C?
13.2
In Problem 13.1, the adjustment had nine degrees of freedom.
(a) Did the adjustment pass the 2 test at a 95% confidence level?
(b) Assuming it passed the 2 test in part (a), what are the estimated
errors in the adjusted parameters?
For Problems 13.3 to 13.8, determine the estimated errors in the adjusted
elevations.
13.3
Problem 12.1
13.4
Problem 12.4
13.5
Problem 12.7
13.6
Problem 12.9
13.7
Problem 12.11
13.8
Problem 12.13
For each problem, calculate the estimated errors for the adjusted elevation
differences.
13.9
Problem 12.1
13.10 Problem 12.4
13.11 Problem 12.7
13.12 Problem 12.9
13.13 Problem 12.11
13.14 Calculate the adjusted length AD and its estimated error given Figure
P13.14 and observational data below (assume equal weights):
l1 ⫽ 100.01
l4 ⫽ 99.94
l2 ⫽ 200.00
l5 ⫽ 200.02
l3 ⫽ 300.02
l6 ⫽ 299.98
PROBLEMS
231
Figure P13.14
13.15 Use Figure P13.15 and the data below to answer the following
questions.
Elevation of BM A ⴝ 263.453 m
Obs
1
2
3
4
5
6
7
From
To
BM
BM
V
V
BM
BM
Y
V
V
X
Y
Y
X
X
A
B
B
A
⌬Elev (m)
25.102
(m)
10.987
24.606
17.993
36.085
⫺13.295
Ⳳ0.018
Ⳳ0.019
Ⳳ0.016
Ⳳ0.021
Ⳳ0.017
Ⳳ0.021
Ⳳ0.018
⫺6.287
Elevation of BM B ⴝ 294.837 m
Obs
From
To
⌬Elev (m)
(m)
8
9
10
11
12
13
14
Y
W
V
BM A
BM B
W
X
Z
Z
W
W
X
X
Z
⫺20.732
Ⳳ0.022
Ⳳ0.022
Ⳳ0.021
Ⳳ0.017
Ⳳ0.020
Ⳳ0.018
Ⳳ0.020
18.455
⫺14.896
10.218
4.693
25.883
⫺7.456
What is:
(a) the most probable elevation for each of stations V, W, X, Y, and
Z?
(b) the estimated error in each elevation?
(c) the estimated error in each adjusted observation?
(d) the estimated error in the elevation difference from benchmark A
to station Z?
232
PRECISION OF INDIRECTLY DETERMINED QUANTITIES
Figure P13.15
13.16 Do a 2 test in Problem 13.15. What observation might contain a
blunder?
13.17 Repeat Problem 13.15 without observations 3, 4, and 10.
13.18 Repeat Problem 13.15 without observations 4, 8, 9, and 12.
13.19 Use ADJUST to do Problems 13.15, 13.17, and 13.18. Explain any
differences in the adjustment results.
Programming Problems
13.20 Adapt the program developed in Problem 12.17 to compute and tabulate the adjusted:
(a) elevations and their estimated errors.
(b) elevation differences and their estimated errors.
13.21 Adapt the program developed in Problem 12.18 to compute and tabulate the adjusted:
(a) elevations and their estimated errors.
(b) elevation differences and their estimated errors.
CHAPTER 14
ADJUSTMENT OF HORIZONTAL
SURVEYS: TRILATERATION
14.1
INTRODUCTION
Horizontal surveys are performed for the purpose of determining precise relative horizontal positions of points. They have traditionally been accomplished by trilateration, triangulation, and traverse. These traditional types of
surveys involve making distance, direction, and angle observations. As with
all types of surveys, errors will occur in making these measurements, and
thus they must be analyzed and if acceptable, adjusted. In the following three
chapters, procedures are described for adjusting trilateration, triangulation,
and traverse surveys, in that order.
In recent years, the global positioning system (GPS) has gradually been
replacing these traditional procedures for conducting precise horizontal control surveys. In fact, GPS not only yields horizontal positions, but it gives
ellipsoidal heights as well. Thus, GPS provides three-dimensional surveys.
Again as with all observations, GPS observations contain errors and must be
adjusted. In Chapter 17 we discuss the subject of GPS surveying in more
detail and illustrate methods for adjusting networks surveyed by this
procedure.
Horizontal surveys, especially those covering a large extent, must account
for the systematic effects of the Earth’s curvature. One way this can be accomplished is to do the computations using coordinates from a mathematically rigorous map projection system such as the state plane system or a local
plane coordinate system that accounts rigorously for the Earth’s curvature.
Map projection coordinate systems are presented in Appendix F. In the following chapters, methods are developed for adjusting horizontal surveys using
parametric equations that are based on plane coordinates. In Chapter 23, a
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
233
234
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
three-dimensional geodetic network adjustment is developed for traditional
surveying observations, including differential leveling, slant distances, and
vertical angles.
It should be noted that if state plane coordinates are used, the numbers are
usually rather large. Consequently, when they are used in mathematical computations, errors due to round-off and truncation can occur. This can be prevented by translating the origin of the coordinates prior to adjustment, a
process that simply involves subtracting a constant value from all coordinates.
Then after the adjustment is finished, the true origin is restored by adding the
constants to the adjusted values. This procedure is demonstrated with the
following example.
Example 14.1 Assume that the NAD 83 state plane coordinates of three
control stations to be used in a horizontal survey adjustment are as given
below. Translate the origin.
Point
Easting (m)
Northing (m)
A
B
C
698,257.171
698,734.839
698,866.717
172,068.220
171,312.344
170,696.617
SOLUTION
Step 1: Many surveyors prefer to work in feet, and some jobs require it.
Thus, in this step the eastings and northings, respectively, are converted to
X and Y values in feet by multiplying by 3.28083333. This is the factor
for converting meters to U.S. survey feet and is based on there being
exactly 39.37 inches per meter. After making the multiplications, the coordinates in feet are as follows:
Point
X (ft)
Y (ft)
A
B
C
2,290,865.40
2,292,432.55
2,292,865.22
564,527.15
562,047.25
560,027.15
Step 2: To reduce the sizes of these numbers, an X constant is subtracted
from each X coordinate and a Y constant is subtracted from each Y coordinate. For convenience, these constants are usually rounded to the nearest
thousandth and are normally selected to give the smallest possible coordinates without producing negative values. In this instance, 2,290,000 ft
and 560,000 ft are used as the X and Y constants, respectively. Subtracting
these values from the coordinates yields
14.2
DISTANCE OBSERVATION EQUATION
Point
X⬘ (ft)
Y⬘ (ft)
A
B
C
865.40
2432.55
2865.22
4527.15
2047.25
27.15
235
These X⬘ and Y⬘ coordinates can then be used in the adjustment. After the
adjustment is complete, the coordinates are translated back to their state plane
values by reversing the steps described above, that is, by adding 2,290,000 ft
to all adjusted X coordinates, and adding 560,000 ft to all adjusted Y coordinates. If desired, they can be converted back to meters also.
In the horizontal adjustment problems solved later in this the book, either
translated state plane coordinates or local plane coordinates are used. In this
chapter we concentrate on adjusting trilateration surveys, those involving only
horizontal distance observations. This method of conducting horizontal surveys became common with the introduction of EDM instruments that enable
accurate distance observations to be made rapidly and economically. Trilateration is still possible using today’s modern total station instruments, but as
noted, the procedure is now giving way to traversing and GPS surveys.
14.2
DISTANCE OBSERVATION EQUATION
In adjusting trilateration surveys using the parametric least squares method,
observation equations are written that relate the observed quantities and their
inherent random errors to the most probable values for the x and y coordinates
(the parameters) of the stations involved. Referring to Figure 14.1, the following distance equation can be written for any observation lij :
lij ⫹ vlij ⫽ (xj ⫺ xi)2 ⫹ (yj ⫺ yi)2
Figure 14.1 Observation of a distance.
(14.1)
236
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
In Equation (14.1), lij is the observed distance of a line between stations I
and J, vlij the residual in the observation lij, xi and yi the most probable coordinate values for station I, and xj and yj the most probable coordinate values
for station J. Equation (14.1) is a nonlinear function involving the unknown
variables xi, yi, xj, and yj that can be rewritten as
F(xi,yi,xj,yj) ⫽ lij ⫹ vlij
(14.2)
where
F(xi,yi,xj,yj) ⫽ (xj ⫺ xi)2 ⫹ (yj ⫺ yi)2
As discussed in Section 11.10, a system of nonlinear equations such as
Equation (14.2) can be linearized and solved using a first-order Taylor series
approximation. The linearized form of Equation (14.2) is
F(xi,yi,xj,yj) ⫽ F(xi0,yi0,xj0,yj0) ⫹
⫹
冉冊
冉冊
⭸F
⭸xi
⭸F
⭸xj
dxi ⫹
0
dxj ⫹
0
冉冊
冉冊
⭸F
⭸yi
0
⭸F
⭸yj
0
dyi
dyj
(14.3)
where (⭸F/ ⭸xi)0, (⭸F/ ⭸yi)0, (⭸F/ ⭸xj)0, and (⭸F/ ⭸yj)0 are the partial derivatives
of F with respect to xi, yi xj, and yj, respectively, evaluated with the approximate coordinate values xi0, yi0, xj0, and yj0; xi, yi, xj, and yj are the unknown
parameters; and dxi, dyi, dxj, and dyj are the corrections to the approximation
coordinate values such that
xi ⫽ xi0 ⫹ dxi
yi ⫽ yi0 ⫹ dyi
xj ⫽ xj0 ⫹ dxj
yj ⫽ yj0 ⫹ dyj
(14.4)
The evaluation of partial derivatives is straightforward and will be illustrated with ⭸F/ ⭸xi. Equation (14.2) can be rewritten as
F(xi,yi,xj,yj) ⫽ [(xj ⫺ xi)2 ⫹ (yj ⫺ yi)2]1 / 2
(14.5)
Taking the derivative of Equation (14.5) with respect to xi yields
⭸F
1
⫽ [(xj ⫺ xi)2 ⫹ (yj ⫺ yi)2]⫺1 / 2[2(xj ⫺ xi)(⫺1)]
⭸xi
2
Simplifying Equation (14.6) yields
(14.6)
14.3
TRILATERATION ADJUSTMENT EXAMPLE
⫺(xj ⫺ xi)
xi ⫺ xj
⭸F
⫽
⫽
2
2
⭸xi
IJ
(xj ⫺ xi) ⫹ (yj ⫺ yi)
237
(14.7)
Employing the same procedure, the remaining partial derivatives are
yi ⫺ yj
⭸F
⫽
⭸yi
IJ
xj ⫺ xi
⭸F
⫽
⭸xj
IJ
yj ⫺ yi
⭸F
⫽
⭸yj
IJ
(14.8)
If Equations (14.7) and (14.8) are substituted into Equation (14.3) and the
results substituted into Equation (14.2), the following prototype linearized
distance observation equation is obtained:
冉 冊
冉 冊
yi ⫺ yj
xi ⫺ xj
dxi ⫹
IJ
IJ
0
⫽ klij ⫹ vlij
0
dyi ⫹
冉 冊
xj ⫺ xi
IJ
dxj ⫹
0
冉 冊
yj ⫺ yi
IJ
dyj
0
(14.9)
where (䡠)0 is evaluated at the approximate parameter values, klij ⫽ lij ⫺ IJ0,
and
IJ0 ⫽ F(xi0,yi0,xj0,yj0) ⫽ 兹(xj0 ⫺ xi0)2 ⫹ (yj0 ⫺ yi0)2
14.3
TRILATERATION ADJUSTMENT EXAMPLE
Even though the geometric figures used in trilateration are many and varied,
they are equally adaptable to the observation equation method in a parametric
adjustment. Consider the example shown in Figure 14.2, where the distances
are observed from three stations with known coordinates to a common unknown station U. Since the unknown station has two unknown coordinates
and there are three observations, this results in one redundant observation.
That is, the coordinates of station U could be determined using any two of
Figure 14.2 Trilateration example.
238
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
the three observations. But all three observations can be used simultaneously
and adjusted by the method of least squares to determine the most probable
value for the coordinates of the station.
The observation equations are developed by substituting into prototype
equation (14.9). For example, the equation for distance AU is formed by
interchanging subscript I with A and subscript J with U in Equation (14.9).
In a similar fashion, an equation can be created for each line observed using
the following subscript substitutions:
I
J
A
B
C
U
U
U
When one end of the observed line is a control station, its coordinates are
fixed, and thus those terms can be dropped in prototype equation (14.9).1 This
can be thought of as setting the dx and dy corrections for the control station
equal to zero. In this example, station U always takes the position of J in the
prototype equation, and thus only the coefficients corresponding to dxj and
dyj are used. Using the appropriate substitutions, the following three linearized
observation equations result.
xu0 ⫺ xa
yu ⫺ ya
dxu ⫹ 0
dyu ⫽ (lAU ⫺ AU0) ⫹ vlAU
AU0
AU0
xu0 ⫺ xb
yu ⫺ yb
dxu ⫹ 0
dyu ⫽ (lBU ⫺ AU0) ⫹ vlBU
BU0
BU0
(14.10)
xu0 ⫺ xc
yu ⫺ yc
dxu ⫹ 0
dyu ⫽ (lCU ⫺ CU0) ⫹ vlCU
CU0
CU0
In Equation (14.10),
AU0 ⫽ (xu0 ⫺ xa)2 ⫹ (yu0 ⫺ ya)2
BU0 ⫽ (xu0 ⫺ xb)2 ⫹ (yu0 ⫺ yb)2
CU0 ⫽ (xu0 ⫺ xc)2 ⫹ (yu0 ⫺ yc)2
lau, lbu, and lcu are the observed distances with residuals v; and xu0 and yu0 are
initial coordinate approximations for station U. Equations (14.10) can be expressed in matrix form as
1
The method of dropping control station coordinates from the adjustment, known as elimination
of constraints, is covered in Chapter 20.
14.3
239
TRILATERATION ADJUSTMENT EXAMPLE
(14.11)
JX ⫽ K ⫹ V
where J is the Jacobian matrix of partial derivatives, X the matrix or unknown
corrections dxu and dyu, K the matrix of constants (i.e., the observed lengths,
minus their corresponding lengths computed from the initial approximate coordinates), and V the residual matrix. Equation (14.11) in expanded form is
冤
冥
xu0 ⫺ xa yu0 ⫺ ya
AU0
AU0
xu0 ⫺ xb yu0 ⫺ yb
BU0
BU0
xu0 ⫺ xc
CU0
yu0 ⫺ yc
CU0
冋 册 冤
冥 冤 冥
lAU ⫺ AU0
vlAU
dxu
⫽ lBU ⫺ BU0 ⫹ vlBU
dyu
lCU ⫺ CU0
vlCU
(14.12)
The Jacobian matrix can be formed systematically using the following steps.
Step 1: Head each column with an unknown value.
Step 2: Create a row for every observation.
Step 3: Substitute in the appropriate coefficient corresponding to the column
into each row.
If this procedure is followed for this problem, the Jacobian matrix is
dxu dyu
AU
兩
⭸F ⭸F
dxu dyu
BU
兩
⭸F ⭸F
dxu dyu
CU 兩
⭸F ⭸F
dxu dyu
Once Equation (14.12) is created, the corrections of dxu and dyu, and thus the
most probable coordinate values, xu and yu, can be computed using Equation
(11.37). Of course, to obtain the final adjusted values, the solution must be
iterated, as discussed in Section 11.10.
Example 14.2 To clarify the computational procedure, a numerical example
for Figure 14.2 is presented. Suppose that the observed distances lAU, lBU, and
lCU are 6049.00, 4736.83, and 5446.49 ft, respectively, and the control stations
have coordinates in units of feet of
240
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
xa ⫽ 865.40
xb ⫽ 2432.55
xc ⫽ 2865.22
ya ⫽ 4527.15
yb ⫽ 2047.25
yc ⫽
27.15
(Note that these are the translated coordinates obtained in Example 14.1.)
Compute the most probable coordinates for station U.
SOLUTION Perform the first iteration.
Step 1: Calculate approximate coordinates for station U.
(a) Calculate azimuth AB from the coordinate values of stations A and B.
x ⫺ xa
AzAB ⫽ tan⫺1 b
⫹ 180⬚
yb ⫺ ya
⫽ tan⫺1
2432.55 ⫺ 865.40
⫹ 180⬚
2047.25 ⫺ 4527.15
⫽ 147⬚42⬘34ⴖ
(b) Calculate the distance between stations A and B from their coordinate
values.
AB ⫽ (xb ⫺ xa)2 ⫹ (xb ⫺ xa)2
⫽ (2432.55 ⫺ 865.20)2 ⫹ (2047.25 ⫺ 4527.15)2
⫽ 2933.58 ft
(c) Calculate azimuth AU0 using the cosine law in triangle AUB:
c2 ⫽ a2 ⫹ b2 ⫺ 2ab cos C
cos(∠UAB) ⫽
6049.002 ⫹ 2933.582 ⫺ 4736.832
2(6049.00)(2933.58)
∠UAB ⫽ 50⬚06⬘50ⴖ
AzAU0 ⫽ 147⬚42⬘34ⴖ ⫺ 50⬚06⬘50ⴖ ⫽ 97⬚35⬘44ⴖ
(d) Calculate the coordinates for station U.
xu0 ⫽ 865.40 ⫹ 6049.00 sin 97⬚35⬘44ⴖ ⫽ 6861.325 ft
yu0 ⫽ 4527.15 ⫹ 6049.00 cos 97⬚35⬘44ⴖ ⫽ 3727.596 ft
14.3
TRILATERATION ADJUSTMENT EXAMPLE
241
Step 2: Calculate AU0, BU0, and CU0. For this first iteration, AU0 and BU0
are exactly equal to their respective observed distances since xu0 and yu0
were calculated using these quantities. Thus,
AU0 ⫽ 6049.00
BU0 ⫽ 4736.83
CU0 ⫽ (6861.325 ⫺ 2865.22)2 ⫹ (3727.596 ⫺ 27.15)2 ⫽ 5446.298 ft
Step 3: Formulate the matrices.
(a) The elements of the Jacobian matrix in Equation (14.12) are2
j11 ⫽
6861.325 ⫺ 865.40
⫽ 0.991
6049.00
j12 ⫽
3727.596 ⫺ 4527.15
⫽ ⫺0.132
6049.00
j21 ⫽
6861.325 ⫺ 2432.55
⫽ 0.935
4736.83
j22 ⫽
3727.596 ⫺ 2047.25
⫽ 0.355
4736.83
j31 ⫽
6861.325 ⫺ 2865.22
⫽ 0.734
5446.298
j32 ⫽
3727.596 ⫺ 27.15
⫽ 0.679
5446.298
(b) The elements of the K matrix in Equation (14.12) are
k1 ⫽ 6049.00 ⫺ 6049.00 ⫽ 0.000
k2 ⫽ 4736.83 ⫺ 4736.83 ⫽ 0.000
k3 ⫽ 5446.49 ⫺ 5446.298 ⫽ 0.192
Step 4: The matrix solution using Equation (11.37) is
X ⫽ (J TJ)⫺1J TK
J TJ ⫽
冋
册冤
冋
0.991 0.935 0.734
⫺0.132 0.355 0.679
(J TJ)⫺1 ⫽
冥 冋
册
0.991 ⫺0.132
2.395 0.699
0.935
0.355 ⫽
0.699 0.605
0.735
0.679
册
1
0.605 ⫺0.699
⫺
0.699
2.395
0.960
2
Note that the denominators in the coefficients of step 3a are distances computed from the approximate coordinates. Only the distances computed for the first iteration will match the measured
distances exactly. Do not use measured distances for the denominators of these coefficients.
242
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
冋
X⫽
册冤 冥 冋 册
册冋 册 冋 册
0.000
0.141
0.000 ⫽
0.130
0.192
0.991 0.935 0.734
⫺0.132 0.355 0.679
J TK ⫽
冋
1
0.605 ⫺0.699
⫺
0.699
2.395
0.960
0.141
⫺0.006
⫽
0.130
0.222
The revised coordinates of U are
xu ⫽ 6861.325 ⫺ 0.006 ⫽ 6861.319
yu ⫽ 3727.596 ⫹ 0.222 ⫽ 3727.818
Now perform the second iteration.
Step 1: Calculate AU0, BU0, and CU0.
AU0 ⫽ 兹(6861.319 ⫺ 865.40)2 ⫹ (3727.818 ⫺ 4527.15)2 ⫽ 6048.965 ft
BU0 ⫽ 兹(6861.319 ⫺ 2432.55)2 ⫹ (3727.818 ⫺ 2047.25)2 ⫽ 4736.909 ft
CU0 ⫽ 兹(6861.319 ⫺ 2865.22)2 ⫹ (3727.818 ⫺ 27.15)2 ⫽ 5446.444 ft
Notice that these computed distances no longer match their observed
counterparts.
Step 2: Formulate the matrices. With these minor changes in the lengths, the
J matrix (to three places) does not change, and thus (JTJ)⫺1 does not change
either. However, the K matrix does change, as shown by the following
computations.
k1 ⫽ 6049.00 ⫺ 6048.965 ⫽ 0.035
k2 ⫽ 4736.83 ⫺ 4736.909 ⫽ ⫺0.079
k3 ⫽ 5446.49 ⫺ 5446.444 ⫽ 0.046
Step 3: Matrix solution
J TK ⫽
X⫽
冋
册冤
册冋
0.991 0.935 0.734
⫺0.132 0.355 0.679
冋
1
0.605 ⫺0.699
2.395
0.960 ⫺0.699
冥 冋
册 冋
0.035
⫺0.079
⫽
0.046
⫺0.005
⫺0.001
⫽
册
册
⫺0.005
⫺0.001
⫺0.002
0.001
14.4
FORMULATION OF A GENERALIZED COEFFICIENT MATRIX
243
The revised coordinates of U are
xu ⫽ 6861.319 ⫺ 0.002 ⫽ 6861.317
yu ⫽ 3727.818 ⫹ 0.001 ⫽ 3727.819
Satisfactory convergence is shown by the very small corrections in the
second iteration. This problem has also been solved using the program ADJUST. Values computed include the most probable coordinates for station U,
their standard deviations, the adjusted lengths of the observed distances, their
residuals and standard deviations, and the reference variance and reference
standard deviation. These are tabulated as shown below.
*****************
Adjusted stations
*****************
Station
X
Y
Sx
Sy
========================================================
U
6,861.32
3,727.82
0.078
0.154
*******************************
Adjusted Distance Observations
*******************************
Station
Station
Occupied
Sighted
Distance
V
S
========================================================
A
U
6,048.96
⫺0.037
0.090
B
U
4,736.91
0.077
0.060
C
U
5,446.44
⫺0.047
0.085
Adjustment Statistics
S20 ⫽ 0.00954
S0 ⫽ Ⳳ0.10
14.4 FORMULATION OF A GENERALIZED COEFFICIENT MATRIX
FOR A MORE COMPLEX NETWORK
In the trilaterated network of Figure 14.3, all lines were observed. Assume
that stations A and C are control stations. For this network, there are 10
observations and eight unknowns. Stations A and C can be fixed by giving
the terms dxa, dya, dxc, and dyc zero coefficients, which effectively drops these
244
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
Figure 14.3 Trilateration network.
terms from the solution. The coefficient matrix formulated from prototype
equation (14.9) has nonzero elements, as indicated in Table 14.1. In this table
the appropriate coefficient from Equation (14.9) is indicated by its corresponding unknown terms of dxi, dyi, dxj, or dyj.
14.5
COMPUTER SOLUTION OF A TRILATERATED QUADRILATERAL
The quadrilateral shown in Figure 14.4 was adjusted using the software
MATRIX. In this problem, points Bucky and Badger are control stations
whose coordinates are held fixed. The five distances observed are:
Line
Distance (ft)
Badger–Wisconsin
Badger–Campus
Wisconsin–Campus
Wisconsin–Bucky
Campus–Bucky
5870.302
7297.588
3616.434
5742.878
5123.760
TABLE 14.1 Structure of the Normal Matrix for the Complex Network in
Figure 14.3
Unknown
Distance,
IJ
dxb
dyb
dxd
dyd
dxe
dye
dxƒ
dyƒ
AB
AE
BC
BF
BE
CD
CF
DF
DE
EF
dxj
0
dxi
dxi
dxi
0
0
0
0
0
dyj
0
dyi
dyi
dyi
0
0
0
0
0
0
0
0
0
0
dxj
0
dxi
dxi
0
0
0
0
0
0
dyj
0
dyi
dyi
0
0
dxj
0
0
dxj
0
0
0
dxj
dxi
0
dyj
0
0
dyj
0
0
0
dyj
dyi
0
0
0
dxj
0
0
dxj
dxj
0
dxj
0
0
0
dyj
0
0
dyj
dyj
0
dyj
14.5
245
COMPUTER SOLUTION OF A TRILATERATED QUADRILATERAL
Figure 14.4 Quadrilateral network.
The state plane control coordinates in units of feet for station Badger are x
⫽ 2,410,000.000 and y ⫽ 390,000.000, and for Bucky are x ⫽ 2,411,820.000
and y ⫽ 386,881.222.
Step 1: To solve this problem, approximate coordinates are first computed
for stations Wisconsin and Campus. This is done following the procedures
used in Section 14.3, with the resulting initial approximations being
Wisconsin:
x ⫽ 2,415,776.819
y ⫽ 391,043.461
Campus:
x ⫽ 2,416,898.227
y ⫽ 387,602.294
Step 2: Following prototype equation (14.9) and the procedures outlined in
Section 14.4, a table of coefficients is established. For the sake of brevity
in Table 14.2, the following station assignments were made: Badger ⫽ 1,
Bucky ⫽ 2, Wisconsin ⫽ 3, and Campus ⫽ 4.
After forming the J matrix, the K matrix is computed. This is done in a
manner similar to step 3 of the first iteration in Example 14.2. The matrices
TABLE 14.2 Structure of the Coefficient or J Matrix for the Example in
Figure 14.4
Badger–Wisconsin
1–3
Badger–Campus
1–4
Wisconsin–Campus
3–4
Wisconsin–Bucky
3–2
Campus–Bucky
4–2
dxWisconsin
dyWisconsin
dxCampus
dyCampus
x30 ⫺ x1
(1–3)0
y30 ⫺ y1
(1–3)0
0
0
0
0
x30 ⫺ x40
(3–4)0
x30 ⫺ x20
(3–2)0
y30 ⫺ y40
(3–4)0
y30 ⫺ y20
(3–2)0
x40 ⫺ x1
(1–4)0
x40 ⫺ x30
(3–4)0
y40 ⫺ y1
(1–4)0
y40 ⫺ y30
(3–4)0
0
0
0
0
x40 ⫺ x20
(4–2)0
y40 ⫺ y20
(4–2)0
246
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
were entered into a file following the formats listed in the Help file for program MATRIX. Following are the input data, matrices for the first and last
iterations of this three-iteration solution, and the final results tabulated.
*******************************************
Initial approximations for unknown stations
*******************************************
Station
X
Y
=====================================
Wisconsin 2,415,776.819 391,043.461
Campus 2,416,898.227 387,602.294
Control Stations
Station
X
Y
===================================
Badger 2,410,000.000 390,000.000
Bucky 2,411,820.000 386,881.222
*********************
Distance Observations
*********************
Occupied
Sighted
Distance
===============================
Badger Wisconsin 5,870.302
Badger Campus
7,297.588
Wisconsin Campus
3,616.434
Wisconsin Bucky
5,742.878
Campus Bucky
5,123.760
First Iteration Matrices
J Dim: 5x4
K Dim: 5x1 X Dim 4x1
====================================== ========== =========
⫺0.00026
0.084751
0.98408 0.17775 0.00000
0.00000
0.00000 0.00000 0.94457 ⫺0.32832
⫺5.46135 ⫺0.165221
⫺0.30984
0.95079 0.30984 ⫺0.95079
⫺2.84579 ⫺5.531445
0.68900 0.72477 0.00000
0.00000
⫺0.00021
0.959315
0.00000 0.00000 0.99007
0.14058
⫺5.40507 =========
====================================== ==========
JtJ Dim: 4x4
==========================================
0.294595
1.539122
0.379687 ⫺0.096003
0.379687
1.460878
0.294595 ⫺0.903997
⫺0.096003
0.294595
1.968448 ⫺0.465525
0.294595 ⫺0.903997 ⫺0.465525
1.031552
==========================================
14.5
COMPUTER SOLUTION OF A TRILATERATED QUADRILATERAL
247
Inv(N) Dim: 4x4
==========================================
1.198436 ⫺1.160169 ⫺0.099979 ⫺1.404084
⫺1.160169
2.635174
0.194272
2.728324
⫺0.099979
0.194272
0.583337
0.462054
⫺1.404084
2.728324
0.462054
3.969873
==========================================
Final Iteration
J Dim: 5x4
====================================
0.98408 0.17772 0.00000
0.00000
0.00000 0.00000 0.94453 ⫺0.32843
⫺0.30853
0.95121 0.30853 ⫺0.95121
0.68902 0.72474 0.00000
0.00000
0.00000 0.00000 0.99002
0.14092
====================================
K Dim:5x1
==========
⫺0.05468
0.07901
⫺0.03675
0.06164
⫺0.06393
==========
X Dim 4x1
=========
0.000627
⫺0.001286
⫺0.000040
0.001814
=========
JtJ Dim: 4x4
==========================================
0.293479
1.538352
0.380777 ⫺0.095191
0.380777
1.461648
0.293479 ⫺0.904809
⫺0.095191
0.293479
1.967465 ⫺0.464182
0.293479 ⫺0.904809 ⫺0.464182
1.032535
==========================================
Qxx ⫽ Inv(N) Dim: 4x4
==========================================
1.198574 ⫺1.160249 ⫺0.099772 ⫺1.402250
⫺1.160249
2.634937
0.193956
2.725964
⫺0.099772
0.193956
0.583150
0.460480
⫺1.402250
2.725964
0.460480
3.962823
==========================================
Qll ⫽ J Qxx Jt Dim: 5x5
======================================================
0.182506 ⫺0.189263
0.838103
0.233921 ⫺0.108806
0.233921
0.662015
0.157210 ⫺0.263698
0.273460
⫺0.108806
0.157210
0.926875
0.122656 ⫺0.127197
0.182506 ⫺0.263698
0.122656
0.794261
0.213356
⫺0.189263
0.273460 ⫺0.127197
0.213356
0.778746
======================================================
248
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
*****************
Adjusted stations
*****************
Station
X
Y
Sx
Sy
=====================================================
Wisconsin 2,415,776.904 391,043.294 0.1488 0.2206
Campus 2,416,892.696 387,603.255 0.1038 0.2705
*******************************
Adjusted Distance Observations
*******************************
Occupied
Sighted
Distance
V
S
===============================================
Badger Wisconsin 5,870.357
0.055 0.1244
Badger
Campus 7,297.509 ⫺0.079 0.1106
Wisconsin
Campus 3,616.471
0.037 0.1308
Wisconsin
Bucky 5,742.816 ⫺0.062 0.1211
Campus
Bucky 5,123.824
0.064 0.1199
-----Reference Standard Deviation
Iterations » 3
⫽ Ⳳ0.135905-----
Notes
1. As noted earlier, it is important that observed distances not be used in
the denominator of the coefficients matrix, J. This is not only theoretically incorrect but can cause slight differences in the final solution, or
even worse, it can cause the system to diverge from any solution! Always compute distances based on the current approximate coordinates.
2. The final portion of the output lists the adjusted x and y coordinates of
the stations, the reference standard deviation, the standard deviations of
the adjusted coordinates, the adjusted line lengths, and their residuals.
3. The Qxx matrix was listed on the last iteration only. It is needed for
calculating the estimated errors of the adjusted coordinates using Equation (14.24) and is also necessary for calculating error ellipses. The
subject of error ellipses is discussed in Chapter 19.
14.6
ITERATION TERMINATION
When programming a nonlinear least squares adjustment, some criteria must
be established to determine the appropriate point at which to stop the iteration
process. Since it is possible to have a set of data that has no solution, it is
also important to determine when that condition occurs. In this section we
14.6
ITERATION TERMINATION
249
describe three methods commonly used to indicate the appropriate time to
end the iteration process.
14.6.1
Method of Maximum Iterations
The simplest procedure of iteration termination involves limiting the number
of iterations to a predetermined maximum. The risk with this method is that
if this maximum is too low, a solution may not be reached at the conclusion
of the process, and if it is too high, time is wasted on unnecessary iterations.
Although this method does not assure convergence, it can prevent the adjustment from continuing indefinitely, which could occur if the solution diverges. When good initial approximations are supplied for the unknown
parameters, a limit of 10 iterations should be well beyond what is required
for a solution since the least squares method converges quadratically.
14.6.2
Maximum Correction
This method was used in earlier examples. It involves monitoring the absolute
size of the corrections. When all corrections become negligibly small, the
iteration process is stopped. The term negligible is relative. For example, if
distances are observed to the nearest foot, it would be foolish to assume that
the size of the corrections will become less than some small fraction of a
foot. Generally, negligible is interpreted as a correction that is less than onehalf the least count of the smallest unit of measure. For instance, if all distances are observed to the nearest 0.01 ft, it would be appropriate to assume
convergence when the absolute size of all corrections is less than 0.005 ft.
Although the solution may continue to converge with continued iterations,
the work to get these corrections is not warranted based on the precision of
the observations.
14.6.3
Monitoring the Adjustment’s Reference Variance
The best method for determining convergence involves monitoring the reference variance and its changes between iterations. Since the least squares
method converges quadratically, the iteration process should definitely be
stopped if the reference variance increases. An increasing reference variance
suggests a diverging solution, which happens when one of two things has
occurred: (1) a large blunder exists in the data set and no solution is possible,
or (2) the maximum correction size is less than the precision of the observations. In the second case, the best solution for the given data set has already
been reached, and when another iteration is attempted, the solution will converge, only to diverge on the next iteration. This apparent bouncing in the
solution is caused by convergence limits being too stringent for the quality
of the data.
250
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
By monitoring the reference variance, convergence and divergence can be
detected. Convergence is assumed when the change in the reference variance
falls below some predefined percentage. Convergence can generally be assumed when the change in the reference variance is less than 1% between
iterations. If the size of the reference variance increases, the solution is diverging and the iteration process should be stopped. It should be noted that
monitoring changes in the reference variance will always show convergence
or divergence in the solution, and thus it is better than any method discussed
previously. However, all methods should be used in concert when doing an
adjustment.
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
14.1
Given the following observed values for the lines in Figure 14.2:
AU ⫽ 2828.83 ft
BU ⫽ 2031.55 ft
CU ⫽ 2549.83 ft
the control coordinates of A, B, and C are:
Station
x (ft)
y (ft)
A
B
C
1418.17
2434.53
3234.86
4747.14
3504.91
2105.56
What are the most probable values for the adjusted coordinates of
station U?
14.2
Do a weighted least squares adjustment using the data in Problem
14.1 with weights base on the following observational errors.
AU ⫽ Ⳳ0.015 ft
BU ⫽ 0.011 ft
CU ⫽ 0.012 ft
(a) What are the most probable values for the adjusted coordinates
of station U?
(b) What is the reference standard deviation of unit weight?
PROBLEMS
251
(c) What are the estimated standard deviations of the adjusted
coordinates?
(d) Tabulate the adjusted distances, their residuals, and the standard
deviations.
Figure 14.3
14.3
Do a least squares adjustment for the following values observed for
the lines in Figure P14.3.
AC ⫽ 2190.04 ft
AD ⫽ 3397.25 ft
BD ⫽ 2250.05 ft
BC ⫽ 2710.38 ft
CD ⫽ 2198.45 ft
In the adjustment, hold the coordinates of stations A and B (in units
of feet) of
xa ⫽ 1423.08 ft
and ya ⫽ 4796.24 ft
xb ⫽ 1776.60 ft
and yb ⫽ 2773.32 ft
(a) What are the most probable values for the adjusted coordinates
of stations C and D?
(b) What is the reference standard deviation of unit weight?
252
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
(c) What are the estimated standard deviations of the adjusted
coordinates?
(d) Tabulate the adjusted distances, their residuals, and the standard
deviations.
14.4
Repeat Problem 14.3 using a weighted least squares adjustment,
where the distance standard deviations are
DA ⫽ Ⳳ0.016 ft
BC ⫽ Ⳳ0.0.018 ft
AC ⫽ Ⳳ0.0.016 ft
BD ⫽ Ⳳ0.0.017 ft
CD ⫽ Ⳳ0.016 ft
14.5
Use the ADJUST software to do Problems 14.3 and 14.4. Explain
any differences in the adjustments.
14.6
Using the trilaterated Figure 14.3 and the data below, compute the
most probable station coordinates and their standard deviations.
Initial approximations of stations
Control stations
Station Easting (m) Northing (m)
Station Easting (m) Northing (m)
B
D
E
F
12,349.500
17,927.677
13,674.750
14,696.838
14,708.750
11,399.956
10,195.970
12,292.118
A
C
10,487.220
16,723.691
11,547.206
14,258.338
Distance observations
14.7
Station
Occupied
Station
Sighted
Distance (m)
(m)
A
B
C
D
E
B
B
F
F
F
B
C
D
E
A
E
F
E
C
D
3669.240
4397.254
3101.625
4420.055
3462.076
4703.319
3369.030
2332.063
2823.857
3351.737
0.013
0.015
0.012
0.015
0.013
0.016
0.012
0.010
0.011
0.012
Using the station coordinates and trilateration data given below, find:
(a) the most probable coordinates for station E.
(b) the reference standard deviation of unit weight.
(c) the standard deviations of the adjusted coordinates.
(d) the adjusted distances, their residuals, and the standard deviations.
PROBLEMS
Control stations
253
Initial approximations
Station
Easting
(m)
Northing
(m)
A
B
C
D
100,643.154
101,093.916
137,515.536
139,408.739
38,213.066
67,422.484
67,061.874
37,544.403
Station
E
Easting
(m)
Northing
(m)
119,665,336 53,809.452
Distance observations
14.8
From
To
Distance (m)
S (m)
A
B
C
D
E
E
E
E
24,598.543
23,026.189
22,231.945
25,613.764
0.074
0.069
0.067
0.077
Using the station coordinates and trilateration data given below, find:
(a) the most probable coordinates for the unknown stations.
(b) the reference standard deviation of unit weight.
(c) the standard deviations of the adjusted coordinates.
(d) the adjusted distances, their residuals, and the standard deviations.
Initial approximations
Control station
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
A
D
92,890.04
93,971.87
28,566.74
80,314.29
B
C
E
F
G
H
I
J
K
93,611.26
93,881.71
111,191.00
110,109.17
110,019.02
131,475.32
130,213.18
129,311.66
128,590.44
47,408.62
64,955.36
38,032.76
57,145.10
73,102.09
28,837.20
46,777.56
64,717.91
79,142.31
Distance observations
From
To
Distance (ft)
S (ft)
A
B
C
D
C
C
B
B
C
D
G
G
F
F
18,855.74
17,548.79
15,359.17
17,593.38
18,077.20
18,009.22
19,156.82
0.06
0.05
0.05
0.05
0.06
0.06
0.06
254
ADJUSTMENT OF HORIZONTAL SURVEYS: TRILATERATION
B
A
H
I
H
I
J
I
J
J
K
E
F
14.9
E
E
E
E
I
J
K
F
F
G
G
F
G
19,923.71
20,604.19
22,271.36
20,935.94
17,984.75
17,962.99
14,442.41
22,619.85
20,641.79
21,035.82
19,529.02
19,142.85
15,957.22
0.06
0.06
0.07
0.06
0.06
0.06
0.05
0.07
0.06
0.06
0.06
0.06
0.05
Use the ADJUST software to do Problem 14.5.
14.10 Use the ADJUST software to do Problem 14.8.
14.11 Describe the methods used to detect convergence in a nonlinear least
squares adjustment and the advantages and disadvantages of each.
Programming Problems
14.12 Create a computational program that computes the distance, coefficients, and klij in Equation (14.9) between stations I and J given their
initial coordinate values. Use this spreadsheet to determine the matrix
values necessary for solving Problem 14.5.
14.13 Create a computational program that reads a data file containing station coordinates and distances and generates the J, W, and K matrices,
which can be used by the MATRIX program. Demonstrate that this
program works by using the data of Problem 14.5.
14.14 Create a computational program that reads a file containing the J, W,
and K matrices and finds the most probable value for the station
coordinates, the reference standard deviation, and the standard deviations of the station coordinates. Demonstrate that this program works
by solving Problem 14.5.
14.15 Create a computational program that reads a file containing control
station coordinates, initial approximations of unknown stations, and
distance observations. The program should generate the appropriate
matrices for a least squares adjustment, do the adjustment, and print
out the final adjusted coordinates, their standard deviations, the final
adjusted distances, their residuals, and the standard deviations in the
adjusted distances. Demonstrate that this program works by solving
Problem 14.5.
CHAPTER 15
ADJUSTMENT OF HORIZONTAL
SURVEYS: TRIANGULATION
15.1
INTRODUCTION
Prior to the development of electronic distance measuring equipment and the
global positioning system, triangulation was the preferred method for extending horizontal control over long distances. The positions of widely spaced
stations were computed from measured angles and a minimal number of measured distances called baselines. This method was used extensively by the
National Geodetic Survey in extending much of the national network. Triangulation is still used by many surveyors in establishing horizontal control,
although surveys that combine trilateration (distance observations) with triangulation (angle observations) are more common. In this chapter, methods
are described for adjusting triangulation networks using least squares.
A least squares triangulation adjustment can use condition equations or
observation equations written in terms of either azimuths or angles. In this
chapter the observation equation method is presented. The procedure involves
a parametric adjustment where the parameters are coordinates in a plane rectangular system such as state plane coordinates. In the examples, the specific
types of triangulations known as intersections, resections, and quadrilaterals
are adjusted.
15.2
AZIMUTH OBSERVATION EQUATION
The azimuth equation in parametric form is
azimuth ⫽ ␣ ⫹ C
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
(15.1)
255
256
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Figure 15.1 Relationship between the azimuth and the computed angle, ␣.
where ␣ ⫽ tan⫺1[(xj ⫺ xi)/(yj ⫺ yi)]; xi and yi are the coordinates of the
occupied station I; xj and yj are the coordinates of the sighted station J; and
C is a constant that depends on the quadrant in which point J lies, as shown
in Figure 15.1.
From the figure, Table 15.1 can be constructed, which relates the algebraic
sign of the computed angle ␣ in Equation (15.1) to the value of C and the
value of the azimuth.
15.2.1
Linearization of the Azimuth Observation Equation
Referring to Equation (15.1), the complete observation equation for an observed azimuth of line IJ is
tan⫺1
xj ⫺ xi
⫹ C ⫽ Azij ⫹ vAzij
yj ⫺ yi
(15.2)
where Azij is the observed azimuth, vAzij the residual in the observed azimuth,
xi and yi the most probable values for the coordinates of station I, xj and yj
the most probable values for the coordinates of station J, and C a constant
with a value based on Table 15.1. Equation (15.2) is a nonlinear function
involving variables xi, yi, xj, and yj, that can be rewritten as
F(xi,yi,xj,yj) ⫽ Azij ⫹ vAzij
(15.3)
where
TABLE 15.1 Relationship between the Quadrant, C, and the Azimuth
of the Line
Quadrant
Sign(xj ⫺ xi)
Sign( yj ⫺ yi)
Sign ␣
C
Azimuth
I
II
III
IV
⫹
⫹
⫺
⫺
⫹
⫺
⫺
⫹
⫹
⫺
⫹
⫺
0
180⬚
180⬚
360⬚
␣
␣ ⫹ 180⬚
␣ ⫹ 180⬚
␣ ⫹ 360⬚
15.2
F(xi,yi,xj,yj) ⫽ tan⫺1
257
AZIMUTH OBSERVATION EQUATION
xj ⫺ xi
⫹C
yj ⫺ yi
As discussed in Section 11.10, nonlinear equations such as (15.3) can be
linearized and solved using a first-order Taylor series approximation. The
linearized form of Equation (15.3) is
F(xi,yi,xj,yj) ⫽ F(xi,yi,xj,yj)0 ⫹
⫹
冉冊
冉冊
⭸F
⭸xi
0
⭸F
⭸yj
0
dxi ⫹
冉冊
⭸F
⭸yi
dyi ⫹
0
冉冊
⭸F
⭸xj
dxj
0
(15.4)
dyj
where (⭸F/ ⭸xi)0, (⭸F/ ⭸yi)0, (⭸F/ ⭸xj)0, and (⭸F/ ⭸yj)0 are the partial derivatives
of F with respect to xi, yi, xj, and yj that are evaluated at the initial approximations xi0, yi0, xj0, and yj0, and dxi, dyi, dxj, and dyj are the corrections applied
to the initial approximations after each iteration such that
xi ⫽ xi0 ⫹ dxi yi ⫽ yi0 ⫹ dyi xj ⫽ xj0 ⫹ dxji
yj ⫽ yj0 ⫹ dyi
(15.5)
To determine the partial derivatives of Equation (15.4) requires the prototype
equation for the derivative of tan⫺1u with respect to x, which is
d
1 du
tan⫺1u ⫽
dx
1 ⫹ u2 dx
(15.6)
Using Equation (15.6), the procedure for determining the ⭸F/ ⭸xi is demonstrated as follows:
⭸F
1
⫺1
⫽
⭸xi
1 ⫹ [(xj ⫺ xi)/(yj ⫺ yi)]2 yj ⫺ yi
⫽
⫺1(yj ⫺ yi)
(xj ⫺ xi)2 ⫹ (yj ⫺ yi)2
⫽
yi ⫺ yj
IJ 2
(15.7)
By employing the same procedure, the remaining partial derivatives are
xj ⫺ xi
⭸F
⫽
⭸yi
IJ 2
xj ⫺ yi
⭸F
⫽
⭸xj
IJ 2
where IJ 2 ⫽ (xj ⫺ xi)2 ⫹ (yj ⫺ yi)2.
xi ⫺ xj
⭸F
⫽
⭸yj
IJ 2
(15.8)
258
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
If Equations (15.7) and (15.8) are substituted into Equation (15.4) and the
results then substituted into Equation (15.3), the following prototype azimuth
equation is obtained:
冉 冊
yi ⫺ yj
IJ 2
dxi ⫹
0
冉 冊
xj ⫺ xi
IJ 2
dyi ⫹
0
冉 冊
yj ⫺ yi
IJ 2
0
dxj ⫹
冉 冊
xi ⫺ xj
IJ 2
dyj
0
(15.9)
⫽ kAzij ⫹ vAzij
Both
冋 冉 冊 册
kAzij ⫽ Azij ⫺ tan⫺1
xj ⫺ xi
yj ⫺ yi
⫹C
and IJ 2 ⫽ (xj ⫺ xi)20 ⫹ (yj ⫺ yi)20
0
are evaluated using the approximate coordinate values of the unknown parameters.
15.3
ANGLE OBSERVATION EQUATION
Figure 15.2 illustrates the geometry for an angle observation. In the figure, B
is the backsight station, F the foresight station, and I the instrument station.
As shown in the figure, an angle observation equation can be written as the
difference between two azimuth observations, and thus for clockwise angles,
∠BIF ⫽ AzIF ⫺ AzIB ⫽ tan⫺1
xƒ ⫺ xi
x ⫺ xi
⫺ tan⫺1 b
⫹ D ⫽ bif ⫹ vbif
yƒ ⫺ yi
yb ⫺ yi
(15.10)
where bif is the observed clockwise angle, vbif the residual in the observed
angle, xb and yb the most probable values for the coordinates of the backsighted station B, xi and yi the most probable values for the coordinates of
Figure 15.2 Relationship between an angle and two azimuths.
15.3
259
ANGLE OBSERVATION EQUATION
the instrument station I, xƒ and yƒ the most probable values for the coordinates
of the foresighted station F, and D a constant that depends on the quadrants
in which the backsight and foresight occur. This term can be computed as
the difference between the C terms from Equation (15.1) as applied to the
backsight and foresight azimuths; that is,
D ⫽ Cif ⫺ Cib
Equation (15.10) is a nonlinear function of xb, yb, xi, yi, xƒ, and yƒ that can
be rewritten as
F(xb,yb,xi,yi,xƒ,yƒ) ⫽ bif ⫹ vbif
(15.11)
where
F(xb,yb,xi,yi,xƒ,yƒ) ⫽ tan⫺1
xƒ ⫺ xi
x ⫺ xi
⫺ tan⫺1 b
⫹D
yƒ ⫺ yi
yb ⫺ yi
Equation (15.11) expressed as a linearized first-order Taylor series expansion is
冉 冊 冉 冊
冉冊 冉 冊 冉 冊
⭸F
⭸xb
F(xb,yb,xi,yi,xƒ,yƒ) ⫽ F(xb,yb,xi,yi,xƒ,yƒ)0 ⫹
⫹
冉冊
⭸F
⭸xi
dxi ⫹
0
⭸F
⭸yi
⭸F
⭸yb
dxb ⫹
0
dyi ⫹
0
⭸F
⭸xƒ
dyb
0
dxƒ ⫹
0
⭸F
⭸yƒ
dyƒ
0
(15.12)
where ⭸F/ ⭸xb, ⭸F/ ⭸yb, ⭸F/ ⭸xi, ⭸F/ ⭸yi, ⭸F/ ⭸xƒ, and ⭸F/ ⭸yƒ are the partial
derivatives of F with respect to xb, yb, xi, yi, xƒ, and yƒ, respectively.
Evaluating partial derivatives of the function F and substituting into Equation (15.12), then substituting into Equation (15.11), results in the following
equation:
冉
冊
yi ⫺ yb
IB2
⫹
冉
dxb ⫹
0
冊
xb ⫺ xi
IB2
冊
xi ⫺ xb xi ⫺ xƒ
⫺
IB2
IF 2
⫽ kbif ⫹ vbif
where
冉
0
dyb ⫹
0
dyi ⫹
冉
冉
冊
yb ⫺ yi yƒ ⫺ yi
⫺
IB2
IF 2
冊
y ƒ ⫺ yi
IF 2
0
dxƒ ⫹
冉
0
dxi
冊
x i ⫺ xƒ
IF 2
0
dyƒ
(15.13)
260
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
kbif ⫽ bif ⫺ bif0
bif0 ⫽ tan⫺1
IB2 ⫽ (xb ⫺ xi)2 ⫹ (yb ⫺ yi)2
冉
冊
xƒ ⫺ xi
yƒ ⫺ yi
冉
⫺ tan⫺1
0
冊
xb ⫺ xi
yb ⫺ yi
⫹D
0
IF 2 ⫽ (xƒ ⫺ xi)2 ⫹ (yƒ ⫺ yi)2
are evaluated at the approximate values for the unknowns.
In formulating the angle observation equation, remember that I is always
assigned to the instrument station, B to the backsight station, and F to the
foresight station. This station designation must be followed strictly in employing prototype equation (15.13), as demonstrated in the numerical examples that follow.
15.4
ADJUSTMENT OF INTERSECTIONS
When an unknown station is visible from two or more existing control stations, the angle intersection method is one of the simplest and sometimes
most practical ways for determining the horizontal position of a station. For
a unique computation, the method requires observation of at least two horizontal angles from two control points. For example, angles 1, and 2 observed
from control stations R and S in Figure 15.3, will enable a unique computation
for the position of station U. If additional control is available, computations
for the unknown station’s position can be strengthened by observing redundant angles such as angles 3 and 4 in Figure 15.3 and applying the method
of least squares. In that case, for each of the four independent angles, a
linearized observation equation can be written in terms of the two unknown
coordinates, xu and yu.
Example 15.1 Using the method of least squares, compute the most probable coordinates of station U in Figure 15.3 by the least squares intersection
procedure. The following unweighted horizontal angles were observed from
control stations R, S, and T:
Figure 15.3 Intersection example.
15.4
1 ⫽ 50⬚06⬘50ⴖ
2 ⫽ 101⬚30⬘47ⴖ
ADJUSTMENT OF INTERSECTIONS
3 ⫽ 98⬚41⬘17ⴖ
261
4 ⫽ 59⬚17⬘01ⴖ
The coordinates for the control stations R, S, and T are
xr ⫽ 865.40
xs ⫽ 2432.55
xt ⫽ 2865.22
yr ⫽ 4527.15
ys ⫽ 2047.25
yt ⫽ 27.15
SOLUTION
Step 1: Determine initial approximations for the coordinates of station U.
(a) Using the coordinates of stations R and S, the distance RS is computed
as
RS ⫽ 兹(2432.55 ⫺ 865.40)2 ⫹ (4527.15 ⫺ 2047.25)2 ⫽ 2933.58 ft
(b) From the coordinates of stations R and S, the azimuth of the line
between R and S can be determined using Equation (15.2). Then the
initial azimuth of line RU can be computed by subtracting 1 from the
azimuth of line RS:
AzRS ⫽ tan⫺1
xs ⫺ xr
865.40 ⫺ 2432.55
⫹ C ⫽ tan⫺1
⫹ 180⬚
ys ⫺ yr
4527.15 ⫺ 2047.25
⫽ 147⬚ ⫹ 42⬘34ⴖ
AzRU0 ⫽ 147⬚42⬘34ⴖ ⫺ 50⬚06⬘50ⴖ ⫽ 97⬚35⬘44ⴖ
(c) Using the sine law with triangle RUS, an initial length for RU0 can be
calculated as
RU0 ⫽
RS sin 2
2933.58 sin(100⬚30⬘47ⴖ)
⫽
⫽ 6049.00 ft
sin(180⬚ ⫺ 1 ⫺ 2)
sin(28⬚27⬘23ⴖ)
(d) Using the azimuth and distance for RU0 computed in steps 1(b) and
1(c), initial coordinates for station U are computed as
xu0 ⫽ xr ⫹ RU0 sin AzRU0 ⫽ 865.40 ⫹ 6049.00 sin(97⬚35⬘44ⴖ)
⫽ 6861.35
yu0 ⫽ yr ⫹ RU0 cos AzRU0 ⫽ 865.40 ⫹ 6049.00 cos(97⬚35⬘44ⴖ)
⫽ 3727.59
262
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
(e) Using the appropriate coordinates, the initial distances for SU and TU
are calculated as
SU0 ⫽ 兹(6861.35 ⫺ 2432.55)2 ⫹ (3727.59 ⫺ 2047.25)2
⫽ 4736.83 ft
TU0 ⫽ 兹(6861.35 ⫺ 2865.22)2 ⫹ (3727.59 ⫺ 27.15)2
⫽ 5446.29 ft
Step 2: Formulate the linearized equations. As in the trilateration adjustment,
control station coordinates are held fixed during the adjustment by assigning zeros to their dx and dy values. Thus, these terms drop out of prototype
equation (15.13). In forming the observation equations, b, i, and ƒ are
assigned to the backsight, instrument, and foresight stations, respectively,
for each angle. For example, with angle 1, B, I, and F are replaced by U,
R, and S, respectively. By combining the station substitutions shown in
Table 15.2 with prototype equation (15.13), the following observation
equations are written for the four observed angles.
冉
冊
yr ⫺ yu
RU 2
0
冋
tan⫺1
⫽ 1 ⫺
冉
dxu ⫹
冊
ys ⫺ yu
SU 2
tan⫺1
dxu ⫹
0
tan⫺1
⫽ 3 ⫺
yu ⫺ yt
TU 2
⫽ 4 ⫺
dxu ⫹
0
⫽ 2 ⫺
冉
冊
xu ⫺ xr
RU 2
dyu
0
冉
dxu ⫹
0
tan⫺1
冊 册
xs ⫺ xr
x ⫺ xr
⫺ tan⫺1 u
ys ⫺ yr
yu ⫺ yr
冉 冊
冋 冉 冊
冊 冉 冊
冋
冉
冊 冉 冊
冋 冉 冊
yu ⫺ ys
SU 2
冉
冉
xs ⫺ xu
SU 2
0
xu ⫺ xs
SU 2
0
⫺ tan⫺1
册
xr ⫺ xs
⫹ 0⬚ ⫹ v2
yr ⫺ ys
dyu
冊
xt ⫺ xs
x ⫺ xs
⫺ tan⫺1 u
yt ⫺ ys
yu ⫺ ys
xu ⫺ xt
yu ⫺ yt
⫹ v1
dyu
0
xu ⫺ xs
yu ⫺ ys
xt ⫺ xu
TU 2
⫹ 0⬚
0
册
⫹ 180⬚
0
⫹ v3
dyu
0
⫺ tan⫺1
0
册
xs ⫺ xt
⫹ 0⬚ ⫹ v4
ys ⫺ yt
(15.14)
15.4
ADJUSTMENT OF INTERSECTIONS
263
TABLE 15.2 Substitutions
Angle
B
I
F
1
2
3
4
U
R
U
S
R
S
S
T
S
U
T
U
Substituting the appropriate values into Equations (15.14) and multiplying
the left side of the equations by to achieve unit consistency,1 the following J and K matrices are formed:
4527.15 ⫺ 3727.59 6861.35 ⫺ 865.40
6049.002
6049.002
3727.59 ⫺ 2047.25 2432.55 ⫺ 6861.35
4.507
33.800
4736.832
4736.832
15.447 ⫺40.713
J⫽
⫽
⫺15.447
40.713
2047.25 ⫺ 3727.59 6861.35 ⫺ 2432.55
25.732 ⫺27.788
4736.832
4736.832
3727.59 ⫺ 27.15 2865.22 ⫺ 6861.35
5446.292
5446.292
冤
冉
冉
冥
冊
冊
冊
冊
50⬚06⬘50ⴖ ⫺ tan⫺1
2432.55 ⫺ 865.40
6861.35 ⫺ 865.40
⫺ tan⫺1
⫹ 0⬚
2047.25 ⫺ 4527.15
3727.59 ⫺ 4527.15
101⬚30⬘47ⴖ ⫺ tan⫺1
6861.35 ⫺ 2432.55
865.40 ⫺ 2432.55
⫺ tan⫺1
⫹ 0⬚
3727.59 ⫺ 2047.25
4527.15 ⫺ 2047.25
冉
K⫽
98⬚41⬘17ⴖ ⫺ tan⫺1
冉
2865.22 ⫺ 2432.55
6861.35 ⫺ 2432.55
⫺ tan⫺1
⫹ 180⬚
27.15 ⫺ 2047.25
3727.59 ⫺ 2047.25
59⬚17⬘01ⴖ ⫺ tan⫺1
6861.35 ⫺ 2865.22
2432.55 ⫺ 2865.22
⫺ tan⫺1
⫹ 0⬚
3727.59 ⫺ 27.15
2047.25 ⫺ 27.15
0.00ⴖ
0.00ⴖ
⫽
⫺0.69ⴖ
⫺20.23ⴖ
冤 冥
1
For these observations to be dimensionally consistent, the elements of the K and V matrices must
be in radian measure, or in other words, the coefficients of the K and J elements must be in the
same units. Since it is most common to work in the sexagesimal system, and since the magnitudes
of the angle residuals are generally in the range of seconds, the units of the equations are converted
to seconds by (1) multiplying the coefficients in the equation by , which is the number of seconds
per radian, or 206,264.8ⴖ / rad, and (2) computing the K elements in units of seconds.
264
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Notice that the initial coordinates for xu0 and yu0 were calculated using 1
and 2, and thus their K-matrix values are zero for the first iteration. These
values will change in subsequent iterations.
Step 3: Matrix solution. The least squares solution is found by applying
Equation (11.37).
J TJ ⫽
冋
Qxx ⫽ (J TJ)⫺1 ⫽
J TK ⫽
册
1159.7 ⫺1820.5
5229.7
⫺1820.5
冋
冋
册
0.001901 0.000662
0.000662 0.000422
册
⫺509.9
534.1
X ⫽ (J TJ)⫺1(J TK) ⫽
dxu ⫽ ⫺0.62 ft
冋
册冋
0.001901 0.000662
0.000662 0.000422
册 冋 册
⫺509.9
534.1
⫽
dxu
dyu
and dyu ⫽ ⫺0.11 ft
Step 4: Add the corrections to the initial coordinates for station U:
xu ⫽ xu0 ⫹ dxu ⫽ 6861.35 ⫺ 0.62 ⫽ 6860.73
(15.15)
yu ⫽ yu0 ⫹ dyu ⫽ 3727.59 ⫺ 0.11 ⫽ 3727.48
Step 5: Repeat steps 2 through 4 until negligible corrections occur. The next
iteration produced negligible corrections, and thus Equations (15.15) produced the final adjusted coordinates for station U.
Step 6: Compute post-adjustment statistics. The residuals for the angles are
冤
冥冋
4.507
33.80
15.447 ⫺40.713
V ⫽ JX ⫺ K ⫽
⫺15.447
40.713
25.732 ⫺27.788
⫽
册
⫺0.62
⫺0.11
冤 冥
0.00ⴖ
0.00ⴖ
⫺
⫺0.69ⴖ
⫺20.23ⴖ
冤 冥
⫺6.5ⴖ
⫺5.1ⴖ
5.8ⴖ
7.3ⴖ
The reference variance (standard deviation of unit weight) for the adjustment is computed using Equation (12.14) as
15.5
T
V V ⫽ [⫺6.5 ⫺5.1
S0 ⫽
V TV
⫽
m⫺n
5.8
ADJUSTMENT OF RESECTIONS
冤 冥
7.3]
⫺6.5
⫺5.1
5.8
7.3
265
⫽ [155.2]
⫽ 8.8ⴖ
4155.2
⫺2
The estimated errors for the adjusted coordinates of station U, given by
Equation (13.24), are
Sxu ⫽ S0兹Qxuxu ⫽ 8.8兹0.001901 ⫽ 0.38 ft
Syu ⫽ S0兹Qyuyu ⫽ 8.8兹0.000422 ⫽ 0.18 ft
The estimated error in the position of station U is given by
Su ⫽ 兹S 2x ⫹ S 2y ⫽ 兹0.382 ⫹ 0.182 ⫽ 0.42 ft
15.5
ADJUSTMENT OF RESECTIONS
Resection is a method used for determining the unknown horizontal position
of an occupied station by observing a minimum of two horizontal angles to
a minimum of three stations whose horizontal coordinates are known. If more
than three stations are available, redundant observations are obtained and the
position of the unknown occupied station can be computed using the least
squares method. Like intersection, resection is suitable for locating an occasional station and is especially well adapted over inaccessible terrain. This
method is commonly used for orienting total station instruments in locations
favorable for staking projects by radiation using coordinates.
Consider the resection position computation for the occupied station U of
Figure 15.4 having observed the three horizontal angles shown between stations P, Q, R, and S whose positions are known. To determine the position
of station U, two angles could be observed. The third angle provides a check
and allows a least squares solution for computing the coordinates of station U.
Using prototype equation (15.13), a linearized observation equation can be
written for each angle. In this problem, the vertex station is occupied and is
the only unknown station. Thus, all coefficients in the Jacobian matrix follow
the form used for the coefficients of dxi and dyi in prototype equation (15.13).
266
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Figure 15.4 Resection example.
The method of least squares yields corrections, dxu and dyu, which gives the
most probable coordinate values for station U.
15.5.1
Computing Initial Approximations in the Resection Problem
In Figure 15.4 only two angles are necessary to determine the coordinates of
station U. Using stations P, Q, R, and U, a procedure to find the station U’s
approximate coordinate values is
Step 1: Let
∠QPU ⫹ ∠URQ ⫽ G ⫽ 360⬚ ⫺ (∠1 ⫹ ∠2 ⫹ ∠RQP)
(15.16)
Step 2: Using the sine law with triangle PQU yields
QU
PQ
⫽
sin ∠QPU sin ∠1
(a)
QU
QR
⫽
sin ∠URQ sin ∠2
(b)
and with triangle URQ,
Step 3: Solving Equations (a) and (b) for QU and setting the resulting equations equal to each other gives
PQ sin ∠PQU QR sin ∠URQ
⫽
sin ∠1
sin ∠2
(c)
Step 4: From Equation (c), let H be defined as
H⫽
sin ∠QPU QR sin ∠1
⫽
sin ∠URQ PQ sin ∠2
(15.17)
15.5
ADJUSTMENT OF RESECTIONS
267
Step 5: From Equation (15.16),
∠QPU ⫽ G ⫺ ∠URQ
(d)
Step 6: Solving Equation (15.17) for the sin∠QPU, and substituting Equation
(d) into the result gives
sin(G ⫺ ∠URQ) ⫽ H sin∠URQ
(e)
Step 7: From trigonometry
sin(␣ ⫺ ) ⫽ sin ␣ cos  ⫺ cos ␣ sin 
Applying this relationship to Equation (e) yields
sin G ⫺ ∠URQ ⫽ sin G cos∠URQ ⫺ cos G sin∠URQ
(ƒ)
sin G ⫺ ∠URQ ⫽ H sin ∠URQ
(g)
Step 8: Dividing Equation (g) by cos ∠URQ and rearranging yields
sin G ⫽ tan∠URQ[H ⫹ cos(G)]
(h)
Step 9: Solving Equation (h) for ∠URQ gives
∠URQ ⫽ tan⫺1
sin G
H ⫹ cos G
(15.18)
Step 10: From Figure 15.4,
∠RQU ⫽ 180⬚ ⫺ (∠2 ⫹ ∠URQ)
(15.19)
Step 11: Again applying the sine law yields
RU ⫽
QR sin ∠RQU
sin ∠2
(15.20)
Step 12: Finally, the initial coordinates for station U are
xu ⫽ xr ⫹ RU sin(AzRQ ⫺ ∠URQ)
yu ⫽ yr ⫹ RU cos(AzRQ ⫺ ∠URQ)
(15.21)
268
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Example 15.2 The following data are obtained for Figure 15.4. Control
stations P, Q, R, and S have the following (x,y) coordinates: P (1303.599,
1458.615), Q (1636.436, 1310.468), R (1503.395, 888.362), and S (1506.262,
785.061). The observed values for angles 1, 2, and 3 with standard deviations
are as follows:
Backsight
Occupied
Foresight
Angle
S (ⴖ)
P
Q
R
U
U
U
Q
R
S
30⬚29⬘33ⴖ
38⬚30⬘31ⴖ
10⬚29⬘57ⴖ
5
6
6
What are the most probable coordinates of station U?
SOLUTION Using the procedures described in Section 15.5.1, the initial
approximations for the coordinates of station U are:
(a) From Equation (15.10),
∠RQP ⫽ AzPQ ⫺ AzQR ⫽ 293⬚59⬘38.4ⴖ ⫺ 197⬚29⬘38.4ⴖ
⫽ 96⬚30⬘00.0ⴖ
(b) Substituting the appropriate angular values into Equation (15.16) gives
G ⫽ 360⬚ ⫺ (30⬚29⬘33ⴖ ⫹ 38⬚30⬘31ⴖ ⫹ 96⬚30⬘00.0ⴖ) ⫽ 194⬚29⬘56ⴖ
(c) Substituting the appropriate station coordinates into Equation (14.1)
yields
PQ ⫽ 364.318
and QR ⫽ 442.576
(d) Substituting the appropriate values into Equation (15.17) yields H as
H⫽
442.576 sin(30⬚29⬘33ⴖ)
⫽ 0.990027302
364.318 sin(38⬚30⬘31ⴖ)
(e) Substituting previously determined G and H into Equation (15.18),
∠URQ is computed as
∠URQ ⫽ tan⫺1
sin(194⬚29⬘56ⴖ)
⫹ 180⬚
0.990027302 ⫹ cos(194⬚29⬘56ⴖ)
⫽ ⫺85⬚00⬘22ⴖ ⫹ 180⬚ ⫽ 94⬚59⬘36.3ⴖ
15.5
ADJUSTMENT OF RESECTIONS
269
(f) Substituting the value of ∠URQ into Equation (15.19), ∠RQU is determined to be
∠RQU ⫽ 180⬚ ⫺ (38⬚30⬘31ⴖ ⫹ 94⬚59⬘36.3ⴖ) ⫽ 46⬚29⬘52.7ⴖ
(g) From Equation (15.20), RU is
RU ⫽
442.576 sin(46⬚29⬘52.7ⴖ)
⫽ 515.589
sin(38⬚30⬘31ⴖ)
(h) Using Equation (15.1), the azimuth of RQ is
AzRQ ⫽ tan⫺1
1636.436 ⫺ 1503.395
⫹ 0⬚ ⫽ 17⬚29⬘38.4ⴖ
1310.468 ⫺ 888.362
(i) From Figure 15.4, AzRU is computed as
AzRQ ⫽ 197⬚29⬘38.4ⴖ ⫺ 180⬚ ⫽ 17⬚29⬘38.4ⴖ
AzRU ⫽ AzRQ ⫺ ∠URQ ⫽ 360⬚ ⫹ 17⬚29⬘38.4ⴖ ⫺ 94⬚59⬘36.3ⴖ
⫽ 282⬚30⬘02.2ⴖ
(j) Using Equation (15.21), the coordinates for station U are
xu ⫽ 1503.395 ⫹ 515.589 sin AzRU ⫽ 1000.03
yu ⫽ 888.362 ⫹ 515.589 cos AzRU ⫽ 999.96
For this problem, using prototype equation (15.13), the J and K matrices
are
冉
冉
冉
冊 冉
冊 冉
冊 冉
yp ⫺ yu yq ⫺ yu
⫺
UP2
UQ2
J⫽
yq ⫺ yu yr ⫺ yu
⫺
UQ
UR2
yr ⫺ yu ys ⫺ yu
⫺
UR2
US 2
冤
冥
(∠1 ⫺ ∠10)ⴖ
K ⫽ (∠2 ⫺ ∠20)ⴖ
(∠3 ⫺ ∠30)ⴖ
冊
冊
冊
0
xu ⫺ xp xu ⫺ xq
⫺
UP2
UQ2
0
0
xu ⫺ xq xu ⫺ xr
⫺
UQ2
UR2
0
0
xu ⫺ xr xu ⫺ xs
⫺
UR2
US 2
0
270
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Also, the weight matrix W is a diagonal matrix composed of the inverses of
the variances of the angles observed, or
冤 冥
1
52
W⫽ 0
0
0
1
62
0
0
0
1
62
Using the data given for the problem together with the initial approximations computed, numerical values for the matrices were calculated and the
adjustment performed using the program ADJUST. The following results were
obtained after two iterations. The reader is encouraged to adjust these example
problems using both the MATRIX and ADJUST programs supplied.
ITERATION 1
J MATRIX
======================
184.993596
54.807717
214.320813 128.785353
59.963802 ⫺45.336838
K MATRIX
========
⫺0.203359
⫺0.159052
⫺6.792817
X MATRIX
========
⫺0.031107
0.065296
K MATRIX
========
1.974063
⫺1.899346
⫺1.967421
X MATRIX
========
0.000008
0.000004
ITERATION 2
J MATRIX
======================
185.018081
54.771738
214.329904 128.728773
59.943758 ⫺45.340316
INVERSE MATRIX
=======================
0.00116318 ⫺0.00200050
⫺0.00200050
0.00500943
Adjusted stations
Station
X
Y
Sx
Sy
===========================================
U
999.999 1,000.025 0.0206 0.0427
15.6
ADJUSTMENT OF TRIANGULATED QUADRILATERALS
271
Adjusted Angle Observations
Station
Station
Station
Backsighted Occupied Foresighted
Angle
V
S (ⴖ)
========================================================
P
U
Q
30⬚ 29⬘ 31ⴖ ⫺2.0ⴖ 2.3
Q
U
R
38⬚ 30⬘ 33ⴖ 1.9ⴖ 3.1
R
U
S
10⬚ 29⬘ 59ⴖ 2.0ⴖ 3.0
Redundancies
⫽
1
Reference Variance
Reference So
15.6
⫽
0.3636
⫽ 0.60
ADJUSTMENT OF TRIANGULATED QUADRILATERALS
The quadrilateral is the basic figure for triangulation. Procedures like those
used for adjusting intersections and resections are also used to adjust this
figure. In fact, the parametric adjustment using the observation equation
method can be applied to any triangulated geometric figure, regardless of its
shape.
The procedure for adjusting a quadrilateral consists of first using a minimum number of the observed angles to solve the triangles, and computing
initial values for the unknown coordinates. Corrections to these initial coordinates are then calculated by applying the method of least squares. The
procedure is iterated until the solution converges. This yields the most probable coordinate values. A statistical analysis of the results is then made. The
following example illustrates the procedure.
Example 15.3 The following observations are supplied for Figure 15.5. Adjust this figure by the method of unweighted least squares. The observed
angles are as follows:
1 ⫽ 42⬚35⬘29.0⬚ 3 ⫽ 79⬚54⬘42.1ⴖ 5 ⫽ 21⬚29⬘23.9ⴖ 7 ⫽ 31⬚20⬘45.8ⴖ
2 ⫽ 87⬚35⬘10.6ⴖ 4 ⫽ 18⬚28⬘22.4ⴖ 6 ⫽ 39⬚01⬘35.4ⴖ 8 ⫽ 39⬚34⬘27.9ⴖ
The fixed coordinates are
xA ⫽ 9270.33
yA ⫽ 8448.90
xD ⫽ 15,610.58
yD ⫽ 8568.75
272
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Figure 15.5 Quadrilateral.
SOLUTION The coordinates of stations B and C are to be computed in this
adjustment. The Jacobian matrix has the form shown in Table 15.3. The subscripts b, i, and ƒ of the dx’s and dy’s in the table indicate whether stations
B and C are the backsight, instrument, or foresight station in Equation (15.13),
respectively. In developing the coefficient matrix, of course, the appropriate
station coordinate substitutions must be made to obtain each coefficient.
A computer program has been used to form the matrices and solve the
problem. In the program, the angles were entered in the order of 1 through
8. The X matrix has the form
冤冥
dxb
dyb
X⫽
dxc
dyc
The following self-explanatory computer listing gives the solution for this
example. As shown, one iteration was satisfactory to achieve convergence,
since the second iteration produced negligible corrections. Residuals, adjusted
TABLE 15.3 Structure of the Coefficient or J Matrix in Example 15.3
Unknowns
Angle
dxb
dyb
dxc
dyc
1
2
3
4
5
6
7
8
dx(b)
0
dx(i)
dx(i)
0
dx(ƒ)
dx(ƒ)
dy(b)
dy(b)
0
dy(i)
dy(i)
0
dy(ƒ)
dy(ƒ)
dy(b)
dx(ƒ)
dx(b)
dx(b)
0
dx(i)
dx(i)
0
dx(ƒ)
dy(ƒ)
dy(b)
dy(b)
0
dy(i)
dy(i)
0
dy(ƒ)
15.6
ADJUSTMENT OF TRIANGULATED QUADRILATERALS
273
coordinates, their estimated errors, and adjusted angles are tabulated at the
end of the listing.
*******************************************
Initial approximations for unknown stations
*******************************************
Station
X
Y
==============================
B 2,403.600 16,275.400
C 9,649.800 24,803.500
Control Stations
Station
X
Y
==============================
A
9,270.330 8,448.900
D 15,610.580 8,568.750
******************
Angle Observations
******************
Station
Station
Station
Backsighted Occupied Foresighted
Angle
===============================================
B
A
C 42⬚ 35⬘ 29.0ⴖ
C
A
D 87⬚ 35⬘ 10.6ⴖ
C
B
D 79⬚ 54⬘ 42.1ⴖ
D
B
A 18⬚ 28⬘ 22.4ⴖ
D
C
A 21⬚ 29⬘ 23.9ⴖ
A
C
B 39⬚ 01⬘ 35.4ⴖ
A
D
B 31⬚ 20⬘ 45.8ⴖ
B
D
C 39⬚ 34⬘ 27.9ⴖ
Iteration 1
J Matrix
---------------------------------------------⫺14.891521 ⫺13.065362
12.605250
⫺0.292475
0.000000
0.000000 ⫺12.605250
0.292475
20.844399
⫺0.283839 ⫺14.045867
11.934565
8.092990
1.414636
0.000000
0.000000
0.000000
0.000000
1.409396
⫺4.403165
⫺14.045867
11.934565
1.440617 ⫺11.642090
6.798531
11.650726
0.000000
0.000000
⫺6.798531 ⫺11.650726
11.195854
4.110690
K MATRIX
--------⫺1.811949
⫺5.801621
3.508571
1.396963
⫺1.833544
5.806415
⫺5.983393
1.818557
X MATRIX
----------1 ⫺0.011149
2 0.049461
3 0.061882
4 0.036935
274
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
Iteration 2
J Matrix
---------------------------------------------⫺14.891488 ⫺13.065272
12.605219
⫺0.292521
0.000000
0.000000 ⫺12.605219
0.292521
20.844296
⫺0.283922 ⫺14.045752
11.934605
8.092944
1.414588
0.000000
0.000000
0.000000
0.000000
1.409357
⫺4.403162
⫺14.045752
11.934605
1.440533 ⫺11.642083
6.798544
11.650683
0.000000
0.000000
⫺6.798544 ⫺11.650683
11.195862
4.110641
K MATRIX
--------⫺2.100998
⫺5.032381
4.183396
1.417225
⫺1.758129
5.400377
⫺6.483846
1.474357
X MATRIX
----------1 0.000000
2 ⫺0.000000
3 0.000000
4 ⫺0.000001
INVERSE MATRIX
------------------------------0.00700
⫺0.00497
⫺0.00497
0.00762
0.00148
0.01138
0.00160
⫺0.01082
0.00160
0.00148
0.00378
0.00073
⫺0.01082
0.01138
0.00073
0.02365
*****************
Adjusted stations
*****************
Station
X
Y
Sx
Sy
================================================
B
2,403.589
16,275.449
0.4690
0.4895
C
9,649.862
24,803.537
0.3447
0.8622
***************************
Adjusted Angle Observations
***************************
Station
Station
Station
Backsighted
Occupied
Foresighted
Angle
V
S
===============================================================
B
A
C
42⬚ 35⬘ 31.1ⴖ
2.10ⴖ 3.65
C
A
D 87⬚ 35⬘ 15.6ⴖ
5.03ⴖ 4.33
C
B
D 79⬚ 54⬘ 37.9ⴖ ⫺4.18ⴖ 4.29
D
B
A
18⬚ 28⬘ 21.0ⴖ ⫺1.42ⴖ 3.36
D
C
A
21⬚ 29⬘ 25.7ⴖ
1.76ⴖ 3.79
A
C
B
39⬚ 01⬘ 30.0ⴖ ⫺5.40ⴖ 4.37
A
D
B
31⬚ 20⬘ 52.3ⴖ
6.48ⴖ 4.24
B
D
C
39⬚ 34⬘ 26.4ⴖ ⫺1.47ⴖ 3.54
*********************************
Adjustment Statistics
********************************
PROBLEMS
275
Iterations ⫽ 2
Redundancies ⫽ 4
Reference Variance ⫽ 31.42936404
Reference So ⫽ 5.6062
Convergence!
PROBLEMS
15.1
Given the following observations and control station coordinates to
accompany Figure P15.1, what are the most probable coordinates for
station E using an unweighted least squares adjustment?
Figure P15.1
Control stations
Station
X (ft)
Y (ft)
A
B
C
D
10,000.00
11,498.58
12,432.17
11,490.57
10,000.00
10,065.32
11,346.19
12,468.51
Angle observations
15.2
Backsight, b
Occupied, i
Foresight, ƒ
Angle
S (ⴖ)
E
A
E
B
E
C
A
B
B
C
C
D
B
E
C
E
D
E
90⬚59⬘57ⴖ
40⬚26⬘02ⴖ
88⬚08⬘55ⴖ
52⬚45⬘02ⴖ
51⬚09⬘55ⴖ
93⬚13⬘14ⴖ
5.3
4.7
4.9
4.7
4.8
5.0
Repeat Problem 15.1 using a weighted least squares adjustment with
weights of 1/S2 for each angle. What are:
276
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
(a)
(b)
(c)
(d)
15.3
the most probable coordinates for station E?
the reference standard deviation of unit weight?
the standard deviations in the adjusted coordinates for station E?
the adjusted angles, their residuals, and the standard deviations?
Given the following observed angles and control coordinates for the
resection problem of Figure 15.4:
1 ⫽ 49⬚47⬘03ⴖ
2 ⫽ 33⬚21⬘55ⴖ
3 ⫽ 47⬚58⬘53ⴖ
Assuming equally weighted angles, what are the most probable coordinates for station U?
Control stations
Station
X (m)
Y (m)
P
Q
R
S
2423.077
3627.660
3941.898
3099.018
3890.344
3602.291
2728.314
1858.429
15.4
If the estimated standard deviations for the angles in Problem 15.3
are S1 ⫽ 3.1ⴖ, S2 ⫽ 3.0ⴖ, and S3 ⫽ 3.1ⴖ, what are:
(a) the most probable coordinates for station U?
(b) the reference standard deviation of unit weight?
(c) the standard deviations in the adjusted coordinates of station U?
(d) the adjusted angles, their residuals, and the standard deviations?
15.5
Given the following control coordinates and observed angles for an
intersection problem:
Control stations
Station
X (m)
Y (m)
A
B
C
D
100,643.154
101,093.916
137,515.536
139,408.739
38,213.066
67,422.484
67,061.874
37,491.846
Angle observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
D
A
B
C
A
B
C
D
E
E
E
E
319⬚39⬘50⬚
305⬚21⬘17ⴖ
322⬚50⬘35ⴖ
313⬚10⬘22ⴖ
5.0
5.0
5.0
5.0
PROBLEMS
277
What are:
(a) the most probable coordinates for station E?
(b) the reference standard deviation of unit weight?
(c) the standard deviations in the adjusted coordinates of station E?
(d) the adjusted angles, their residuals, and the standard deviations?
15.6
The following control station coordinates, observed angles, and standard deviations apply to the quadrilateral in Figure 15.5.
Initial approximations
Control stations
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
A
D
2546.64
4707.04
1940.26
1952.54
B
C
2243.86
4351.06
3969.72
4010.64
Angle observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
B
C
C
D
D
A
A
B
A
A
B
B
C
C
D
D
C
D
D
A
A
B
B
C
49⬚33⬘30ⴖ
48⬚35⬘54ⴖ
40⬚25⬘44ⴖ
42⬚11⬘56ⴖ
50⬚53⬘07ⴖ
47⬚48⬘47ⴖ
39⬚38⬘34ⴖ
40⬚52⬘20ⴖ
4.2
4.2
4.2
4.2
4.2
4.2
4.2
4.2
Do a weighted adjustment using the standard deviations to calculate
weights. What are:
(a) the most probable coordinates for stations B and C?
(b) the reference standard deviation of unit weight?
(c) the standard deviations in the adjusted coordinates for stations B
and C?
(d) the adjusted angles, their residuals, and the standard deviations?
Figure P15.7
278
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
15.7
For Figure P15.7 and the following observations, perform a weighted
least squares adjustment.
(a) Station coordinate values and standard deviations.
(b) Angles, their residuals, and the standard deviations.
Initial approximations
Control stations
Station
X (m)
Y (m)
Station
X (m)
Y (m)
A
B
114,241.071
116,607.143
91,294.643
108,392.857
C
D
135,982.143
131,567.500
107,857.143
90,669.643
Angle observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
B
C
C
D
D
A
A
B
A
A
B
B
C
C
D
D
C
D
D
A
A
B
B
C
44⬚49⬘15.4ⴖ
39⬚21⬘58.0ⴖ
48⬚14⬘48.9ⴖ
48⬚02⬘49.6ⴖ
38⬚17⬘38.0ⴖ
38⬚53⬘03.9ⴖ
47⬚45⬘56.8ⴖ
54⬚34⬘26.1ⴖ
2.0
2.0
2.0
2.0
2.0
2.0
2.0
2.0
15.8
Do Problem 15.7 using an unweighted least squares adjustment. Compare and discuss any differences or similarities between these results
and those obtained in Problem 15.7.
15.9
The following observations were obtained for the triangulation chain
shown in Figure P15.9.
Initial approximations
Control stations
Station
X (m)
Y (m)
Station
X (m)
Y (m)
A
B
G
H
103,482.143
118,303.571
104,196.429
118,080.357
86,919.643
86,919.643
112,589.286
112,767.857
C
D
E
F
103,616.071
117,991.071
104,375.000
118,169.643
96,116.071
95,580.357
104,196.429
104,598.214
Angle observations
B
I
F
Angle
S (ⴖ)
B
I
F
Angle
S (ⴖ)
C
A
D
B
E
A
B
C
D
C
D
C
B
A
F
58⬚19⬘52ⴖ
32⬚03⬘11ⴖ
29⬚55⬘01ⴖ
61⬚14⬘02ⴖ
54⬚24⬘00ⴖ
3
3
3
3
3
D
C
B
A
F
A
B
C
D
C
B
D
A
C
D
30⬚49⬘56ⴖ
55⬚52⬘51ⴖ
58⬚46⬘53ⴖ
32⬚58⬘06ⴖ
32⬚22⬘05ⴖ
3
3
3
3
3
PROBLEMS
C
D
D
H
E
H
F
D
E
F
E
F
G
H
E
C
C
F
G
F
E
30⬚11⬘27⬚
63⬚02⬘21ⴖ
58⬚37⬘50ⴖ
30⬚21⬘08ⴖ
31⬚25⬘55⬚
30⬚30⬘01ⴖ
58⬚37⬘08ⴖ
3
3
3
3
3
3
3
E
F
C
G
G
F
E
D
E
F
E
F
G
H
F
D
E
H
H
E
G
58⬚48⬘32ⴖ
33⬚59⬘36ⴖ
28⬚34⬘00ⴖ
59⬚11⬘48ⴖ
59⬚36⬘31ⴖ
59⬚01⬘04ⴖ
31⬚17⬘11ⴖ
279
3
3
3
3
3
3
3
Figure P15.9
Use ADJUST to perform a weighted least squares adjustment. Tabulate the final adjusted:
(a) station coordinates and their standard deviations.
(b) angles, their residuals, and the standard deviations.
15.10 Repeat Problem 15.9 using an unweighted least squares adjustment.
Compare and discuss any differences or similarities between these
results and those obtained in Problem 15.9. Use the program ADJUST
in computing the adjustment.
15.11 Using the control coordinates from Problem 14.3 and the following
angle observations, compute the least squares solution and tabulate
the final adjusted:
(a) station coordinates and their standard deviations.
280
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
(b) angles, their residuals, and the standard deviations.
Angle observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
B
C
C
D
D
A
A
B
A
A
B
B
C
C
D
D
C
D
D
A
A
B
B
C
280⬚41⬘06ⴖ
39⬚21⬘53ⴖ
51⬚36⬘16ⴖ
255⬚50⬘03ⴖ
101⬚27⬘17ⴖ
311⬚52⬘38ⴖ
324⬚07⬘04ⴖ
75⬚03⬘50ⴖ
5.2
5.1
5.2
5.2
5.2
5.2
5.1
5.2
15.12 The following observations were obtained for a triangulation chain.
Control stations
Initial approximations
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
A
D
92,890.04
93,971.87
28,566.74
80,314.29
B
C
93,611.26
93,881.71
111,191.00
110,109.17
110,019.02
131,475.32
130,213.18
129,311.66
128,590.44
47,408.62
64,955.36
38,032.76
57,145.10
73,102.09
28,837.20
46,777.56
64,717.91
79,142.31
E
F
G
H
I
J
K
Angle observations
B
I
F
Angle
S (ⴖ)
B
I
F
Angle
S (ⴖ)
B
F
F
D
D
J
C
C
J
E
B
I
E
E
I
G
A
B
C
C
G
G
G
F
F
F
E
E
H
I
J
J
E
E
B
G
K
F
D
G
I
B
F
H
I
F
F
K
60⬚27⬘28ⴖ
58⬚37⬘14ⴖ
65⬚10⬘51ⴖ
62⬚52⬘42ⴖ
137⬚46⬘57ⴖ
66⬚11⬘15ⴖ
50⬚59⬘11ⴖ
63⬚58⬘29ⴖ
48⬚48⬘05ⴖ
62⬚41⬘29ⴖ
58⬚41⬘13ⴖ
49⬚04⬘27ⴖ
61⬚35⬘24ⴖ
51⬚58⬘06ⴖ
71⬚21⬘12ⴖ
63⬚38⬘57ⴖ
2.2
2.1
2.3
2.1
2.4
2.1
2.3
2.1
2.3
2.2
2.5
2.1
2.5
2.1
2.2
2.2
E
C
G
G
K
F
B
G
I
A
F
H
H
F
F
J
B
B
C
D
G
G
F
F
F
E
E
E
I
I
J
K
A
F
F
C
J
C
C
J
E
B
I
A
E
J
G
G
64⬚07⬘06ⴖ
58⬚34⬘09ⴖ
52⬚29⬘14ⴖ
66⬚08⬘08ⴖ
41⬚30⬘18ⴖ
63⬚32⬘13ⴖ
56⬚14⬘56ⴖ
68⬚48⬘05ⴖ
59⬚28⬘49ⴖ
55⬚25⬘19ⴖ
68⬚33⬘06ⴖ
128⬚15⬘52ⴖ
69⬚20⬘10ⴖ
59⬚50⬘35ⴖ
45⬚00⬘39ⴖ
74⬚50⬘46ⴖ
2.1
2.1
2.0
2.2
2.7
2.1
2.6
2.1
2.4
2.3
2.0
2.6
2.0
2.2
2.1
2.3
PROBLEMS
281
Use ADJUST to perform a weighted least squares adjustment. Tabulate the final adjusted:
(a) station coordinates and their standard deviations.
(b) angles, their residuals, and the standard deviations.
15.13 Do Problem 15.12 using an unweighted least squares adjustment.
Compare and discuss any differences or similarities between these
results and those obtained in Problem 15.12. Use the program ADJUST in computing the adjustment.
Use the ADJUST software to do the following problems.
15.14 Problem 15.2
15.15 Problem 15.4
15.16 Problem 15.5
15.17 Problem 15.6
15.18 Problem 15.9
Programming Problems
15.19 Write a computational program that computes the coefficients for prototype equations (15.9) and (15.13) and their k values given the coordinates of the appropriate stations. Use this program to determine
the matrix values necessary to do Problem 15.6.
15.20 Prepare a computational program that reads a file of station coordinates, observed angles, and their standard deviations and then:
(a) writes the data to a file in a formatted fashion.
(b) computes the J, K, and W matrices.
(c) writes the matrices to a file that is compatible with the MATRIX
program.
(d) test this program with Problem 15.6.
15.21 Write a computational program that reads a file containing the J, K,
and W matrices and then:
(a) writes these matrices in a formatted fashion.
(b) performs one iteration of either a weighted or unweighted least
squares adjustment of Problem 15.6.
(c) writes the matrices used to compute the solution and the corrections to the station coordinates in a formatted fashion.
15.22 Write a computational program that reads a file of station coordinates,
observed angles, and their standard deviations and then:
282
ADJUSTMENT OF HORIZONTAL SURVEYS: TRIANGULATION
(a) writes the data to a file in a formatted fashion.
(b) computes the J, K, and W matrices.
(c) performs either a relative or equal weight least squares adjustment
of Problem 15.6.
(d) writes the matrices used to compute the solution and tabulates
the final adjusted station coordinates and their estimated errors
and the adjusted angles, together with their residuals and their
estimated errors.
15.23 Prepare a computational program that solves the resection problem.
Use this program to compute the initial approximations for Problem
15.3.
CHAPTER 16
ADJUSTMENT OF HORIZONTAL
SURVEYS: TRAVERSES
AND NETWORKS
16.1
INTRODUCTION TO TRAVERSE ADJUSTMENTS
Of the many methods that exist for traverse adjustment, the characteristic that
distinguishes the method of least squares from other methods is that distance,
angle, and direction observations are adjusted simultaneously. Furthermore,
the adjusted observations not only satisfy all geometrical conditions for the
traverse but provide the most probable values for the given data set. Additionally, the observations can be rigorously weighted based on their estimated
errors and adjusted accordingly. Given these facts, together with the computational power now provided by computers, it is hard to justify not using least
squares for all traverse adjustment work.
In this chapter we describe methods for making traverse adjustments by
least squares. As was the case in triangulation adjustments, traverses can be
adjusted by least squares using either observation equations or conditional
equations. Again, because of the relative ease with which the equations can
be written and solved, the parametric observation equation approach is discussed.
16.2
OBSERVATION EQUATIONS
When adjusting a traverse using parametric equations, an observation equation
is written for each distance, direction, or angle. The necessary linearized
observation equations developed previously are recalled in the following
equations.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
283
284
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
Distance observation equation:
冉 冊
xi ⫺ xj
IJ
dxi ⫹
0
冉 冊
yi ⫺ yj
IJ
dyi ⫹
0
冉 冊
xj ⫺ xi
IJ
dxj ⫹
0
冉 冊
yj ⫺ yi
IJ
dyj
0
(16.1)
⫽ klij ⫹ vlij
Angle observation equation:
冉
冊
yi ⫺ yb
IB2
⫹
dxb ⫹
0
冉
冉
冊
xb ⫺ xi
IB2
0
冊
xi ⫺ xb xi ⫺ xƒ
⫺
IB2
IF 2
dyb ⫹
dyi ⫹
0
冉
冉
冊
yb ⫺ yi yƒ ⫺ yi
⫺
IB2
IF 2
冊
yƒ ⫺ yi
IF 2
dxƒ ⫹
0
冉
0
dxi
冊
xi ⫺ xƒ
IF 2
dyƒ
(16.2)
0
⫽ kbif ⫹ vbif
Azimuth observation equation:
冉 冊
yi ⫺ yj
IJ 2
dxi ⫹
0
冉 冊
xj ⫺ xi
IJ 2
dyi ⫹
0
⫽ kAzij ⫹ vAzij
冉 冊
yj ⫺ yi
IJ 2
0
dxj ⫹
冉 冊
xi ⫺ xj
IJ 2
dyj
0
(16.3)
The reader should refer to Chapters 14 and 15 to review the specific
notation for these equations. As demonstrated with the examples that follow,
the azimuth equation may or may not be used in traverse adjustments.
16.3
REDUNDANT EQUATIONS
As noted earlier, one observation equation can be written for each angle,
distance, or direction observed in a closed traverse. Thus, if there are n sides
in the traverse, there are n distances and n ⫹ 1 angles, assuming that one
angle exists for orientation of the traverse. For example, each closed traverse
in Figure 16.1 has four sides, four distances, and five angles. Each traverse
also has three points whose positions are unknown, and each point introduces
two unknown coordinates into the solution. Thus, there is a maximum of
2(n ⫺ 1) unknowns for any closed traverse. From the foregoing, no matter
the number of sides, there will always be a minimum of r ⫽ (n ⫹ n ⫹ 1) ⫺
2(n ⫺ 1) ⫽ 3 redundant equations for any closed traverse. That is, every
closed traverse that is fixed in space both positionally and rotationally has a
minimum of three redundant equations.
16.4
NUMERICAL EXAMPLE
285
Figure 16.1 (a) Polygon and (b) link traverses.
16.4
NUMERICAL EXAMPLE
Example 16.1 To illustrate a least squares traverse adjustment, the simple
link traverse shown in Figure 16.2 will be used. The observational data are:
Distance (ft)
RU ⫽ 200.00 0.05
US ⫽ 100.00 0.08
Angle
1 ⫽ 240⬚00⬘ 30ⴖ
2 ⫽ 150⬚00⬘ 30ⴖ
3 ⫽ 240⬚01⬘ 30ⴖ
SOLUTION
Step 1: Calculate initial approximations for the unknown station coordinates.
xu0 ⫽ 1000.00 ⫹ 200.00 sin(60⬚) ⫽ 1173.20 ft
yu0 ⫽ 1000.00 ⫹ 200.00 cos(60⬚) ⫽ 1100.00 ft
Figure 16.2 Simple link traverse.
286
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
Step 2: Formulate the X and K matrices. The traverse in this problem contains
only one unknown station with two unknown coordinates. The elements of
the X matrix thus consist of the dxu and dyu terms. They are the unknown
corrections to be applied to the initial approximations for the coordinates
of station U. The values in the K matrix are derived by subtracting computed quantities, based on initial coordinates, from their respective observed quantities. Note that since the first and third observations were used
to compute initial approximations for station U, their K-matrix values will
be zero in the first iteration.
X⫽
冋 册
dxu
dyu
klRU
200.00 ft
klUS
100.00 ft
K ⫽ k1 ⫽ 240⬚00⬘00ⴖ
k2
150⬚00⬘00ⴖ
k3
240⬚01⬘00ⴖ
⫺
⫺
⫺
⫺
⫺
200.00 ft
0.00 ft
99.81 ft
0.019 ft
240⬚00⬘00ⴖ ⫽
0ⴖ
149⬚55⬘51ⴖ
249ⴖ
240⬚04⬘12ⴖ
⫺192ⴖ
Step 3: Calculate the Jacobian matrix. Since the observation equations are
nonlinear, the Jacobian matrix must be formed to obtain the solution. The
J matrix is formed using prototype equations (16.1) for distances and (16.2)
for angles. As explained in Section 15.4, since the units of the K matrix
that relate to the angles are in seconds, the angle coefficients of the J matrix
must be multiplied by to also obtain units of seconds.
In developing the J matrix using prototype equations (16.1) and (16.2),
subscript substitutions were as shown in Table 16.1. Substitutions of numerical values and computation of the J matrix follow.
J⫽
1173.20 ⫺ 1000.00
200.00
1100.00 ⫺ 1000.00
200.00
1173.20 ⫺ 1223.00
99.81
1100.00 ⫺ 1186.50
99.81
1100.00 ⫺ 1000.00
200.002
1000.00 ⫺ 1173.20
200.002
1000.00 ⫺ 1100.00 1186.50 ⫺ 1100.00
⫺
200.002
99.812
1173.20 ⫺ 1000.00 1173.20 ⫺ 1223.00
⫺
200.002
99.812
1186.50 ⫺ 1100.00
99.812
1173.50 ⫺ 1223.00
99.812
TABLE 16.1 Subscript Substitution
Observation
Subscript Substitution
Length RU
Length US
Angle 1
Angle 2
Angle 3
R ⫽ i, U ⫽ j
U ⫽ i, S ⫽ j
Q ⫽ b, R ⫽ i, U ⫽ ƒ
R ⫽ b, U ⫽ i, S ⫽ ƒ
U ⫽ b, S ⫽ i, T ⫽ ƒ
16.4
J⫽
NUMERICAL EXAMPLE
0.866
0.500
⫺0.499
⫺0.867
515.7
⫺893.2
⫺2306.6
1924.2
1709.9
⫺1031.1
287
Step 4: Formulate the W matrix. The fact that distance and angle observations
have differing observational units and are combined in an adjustment is
resolved by using relative weights that are based on observational variances
in accordance with Equation (10.6). This weighting makes the observation
equations dimensionally consistent. If weights are not used in traverse adjustments (i.e., equal weights are assumed), the least squares problem will
generally either give unreliable results or result in a system of equations
that has no solution. Since weights influence the correction size that each
observation will receive, it is extremely important to use variances that
correspond closely to the observational errors. The error propagation procedures discussed in Chapter 7 aid in the determination of the estimated
errors. Repeating Equation (10.6), the distance and angle weights for this
problem are
distances: wlIJ ⫽
1
1
and angles: wbif ⫽ 2
S 2lIJ
S bif
(16.4)
Again, the units of the weight matrix must match those of the J and K
matrices. That is, the angular weights must be in the same units of measure
(seconds) as the counterparts in the other two matrices. Based on the estimated errors in the observations, the W matrix, which is diagonal, is
1
0.052
(zeros)
1
0.082
W⫽
1
302
1
302
(zeros)
⫽
1
302
400.00
(zeros)
(zeros)
156.2
0.0011
0.0011
0.0011
Step 5: Solve the matrix system. This problem is iterative and was solved
according to Equation (11.39) using the program MATRIX. (Output from
288
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
the solution follows.) The first iteration yielded the following corrections
to the initial coordinates.
dxu ⫽ ⫺0.11 ft
dyu ⫽ ⫺0.01 ft
Note that a second iteration produced zeros for dxu and dyu. The reader is
encouraged to use the MATRIX or ADJUST program to duplicate these
results.
Step 6: Compute the a posteriori adjustment statistics. Also from the program
MATRIX, the residuals and reference standard deviation are
vru ⫽ ⫺0.11 ft
vus ⫽ ⫺0.12 ft
v1 ⫽ ⫺49ⴖ
v2 ⫽ ⫺17ⴖ
v3 ⫽ 6ⴖ
S0 ⫽ 1.82
A 2 test was used as discussed in Section 5.4 to see if the a posteriori
reference variance differed significantly from its a priori value of 1.1 The
test revealed that there was no statistically significant difference between
the a posteriori value of (1.82)2 and its a priori value of 1 at a 99% confidence level, and thus the a priori value should be used for the reference
variance when computing the standard deviations of the coordinates. By
applying Equation (13.24), the estimated errors in the adjusted coordinates
are
SxU ⫽ 1.000.00053 ⫽ 0.023 ft
SyU ⫽ 1.000.000838 ⫽ 0.029 ft
Following are the results from the program ADJUST.
*******************************************
Initial approximations for unknown stations
*******************************************
Station Northing
Easting
===========================
U 1,100.00 1,173.20
1
Since weights are calculated using the formula wi ⫽ 20 / i2, using weights of 1 / i2 implies an a
priori value of 1 for the reference variance (see Chapter 10).
16.4
NUMERICAL EXAMPLE
Control Stations
Station
Easting Northing
===========================
Q 1,000.00
800.00
R 1,000.00 1,000.00
S 1,223.00 1,186.50
T 1,400.00 1,186.50
Distance Observations
Station Station
Occupied Sighted Distance
==================================
R
U
200.00 0.050
U
S
100.00 0.080
Angle Observations
Station
Station
Station
Backsighted Occupied Foresighted
Angle
===================================================
Q
R
U 240⬚ 00⬘ 00ⴖ 30ⴖ
R
U
S 150⬚ 00⬘ 00ⴖ 30ⴖ
U
S
T 240⬚ 01⬘ 00ⴖ 30ⴖ
First Iteration Matrices
J Dim: 5x2
=======================
0.86602
0.50001
⫺0.49894
⫺0.86664
515.68471
⫺893.16591
⫺2306.62893
1924.25287
1790.94422 ⫺1031.08696
=======================
K Dim: 5x1
==========
0.00440
0.18873
2.62001
249.36438
⫺191.98440
==========
X Dim: 2x1
==========
⫺0.11
⫺0.01
==========
W Dim: 5x5
===============================================
400.00000
0.00000 0.00000 0.00000 0.00000
0.00000 156.25000 0.00000 0.00000 0.00000
0.00000
0.00000 0.00111 0.00000 0.00000
0.00000
0.00000 0.00000 0.00111 0.00000
0.00000
0.00000 0.00000 0.00000 0.00111
===============================================
289
290
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
N Dim: 2x2
=========================
10109.947301 ⫺7254.506002
⫺7254.506002
6399.173533
=========================
Final Iteration
J Dim: 5x2
=======================
0.86591
0.50020
⫺0.49972
⫺0.86619
516.14929
⫺893.51028
⫺2304.96717
1925.52297
1788.81788 ⫺1032.01269
=======================
N Dim: 2x2
=========================
10093.552221 ⫺7254.153057
⫺7254.153057
6407.367420
=========================
Qxx Dim: 2x2
==================
0.000530 0.000601
0.000601 0.000838
==================
K Dim: 5x1
==========
0.10723
0.12203
48.62499
17.26820
⫺5.89319
==========
X Dim: 2x1
==========
0.0000
0.0000
==========
Qxx Dim: 2x2
==================
0.000532 0.000602
0.000602 0.000838
==================
J Qxx Jt Dim: 5x5
========================================================
0.001130 ⫺0.001195
⫺0.447052
0.055151
0.391901
⫺0.001195
0.001282
0.510776
⫺0.161934
⫺0.348843
⫺0.447052
0.510776 255.118765 ⫺235.593233 ⫺19.525532
0.055151 ⫺0.161934 ⫺235.593233 586.956593 ⫺351.363360
0.391901 ⫺0.348843 ⫺19.525532 ⫺351.363360 370.888892
========================================================
*****************
Adjusted stations
*****************
Station Northing
Easting
S-N
S-E
=========================================
U 1,099.99 1,173.09 0.029 0.023
16.6
ADJUSTMENT OF NETWORKS
291
*******************************
Adjusted Distance Observations
*******************************
Station Station
Occupied Sighted Distance
V
S
========================================
R
U
199.89 ⫺0.11 0.061
U
S
99.88 ⫺0.12 0.065
***************************
Adjusted Angle Observations
***************************
Station
Station
Station
Backsight Occupied Foresight
Angle
V
S
======================================================
Q
R
U 239⬚ 59⬘ 11ⴖ ⫺49ⴖ 29.0ⴖ
R
U
S 149⬚ 59⬘ 43ⴖ ⫺17ⴖ 44.1ⴖ
S
S
T 240⬚ 01⬘ 06ⴖ
6ⴖ 35.0ⴖ
-----Reference Standard Deviation
Iterations » 2
16.5
⫽ 1.82-----
MINIMUM AMOUNT OF CONTROL
All adjustments require some form of control, and failure to supply a sufficient
amount will result in an indeterminate solution. A traverse requires a minimum of one control station to fix it in position and one line of known direction
to fix it in angular orientation. When a traverse has the minimum amount of
control, it is said to be minimally constrained. It is not possible to adjust a
traverse without this minimum. If minimal constraint is not available, necessary control values can be assumed and the computational process carried
out in arbitrary space. This enables the observed data to be tested for blunders
and errors. In Chapter 21 we discuss minimally constrained adjustments.
A free network adjustment involves using a pseudoinverse to solve systems
that have less than the minimum amount of control. This material is beyond
the scope of this book. Readers interested in this subject should consult Bjerhammar (1973) or White (1987) in the bibliography.
16.6
ADJUSTMENT OF NETWORKS
With the introduction of an EDM instrument, and particularly the total station,
the speed and reliability of making angle and distance observations have in-
292
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
creased greatly. This has led to observational systems that do not conform to
the basic systems of trilateration, triangulation, or traverse. For example, it is
common to collect more than the minimum observations at a station during
a horizontal control survey. This creates what is called a complex network,
referred to more commonly as a network. The least squares solution of a
network is similar to that of a traverse. That is, observation equations are
written for each observation using the prototype equations given in Section
16.2. Coordinate corrections are found using Equation (11.39) and a posteriori
error analysis is carried out.
Example 16.2 A network survey was conducted for the project shown in
Figure 16.3. Station Q has control coordinates of (1000.00, 1000.00) and the
azimuth of line QR is 0⬚06⬘24.5ⴖ with an estimated error of 0.001ⴖ. The
observations and their estimated errors are listed in Table 16.2. Adjust this
survey by the method of least squares.
SOLUTION Using standard traverse coordinate computational methods, the
initial approximations for station coordinates (x,y) were determined to be
R: (1003.07, 2640.00)
S: (2323.07,2638.46)
T: (2661.74, 1096.08)
Under each station heading in the observation columns, a letter representing the appropriate prototype equation dx and dy coefficient appears. For
example, for the first distance QR, station Q is substituted for i in prototype
equation (16.1) and station R replaces j. For the first angle, observed at Q
from R to S, station R takes on the subscript b, Q becomes i, and S is substituted for ƒ in prototype equation (16.2).
Table 16.3 shows the structure of the coefficient matrix for this adjustment
and indicates by subscripts where the nonzero values occur. In this table, the
column headings are the elements of the unknown X matrix dxr, dyr, dxs, dys,
dxt, and dyt. Note that since station Q is a control station, its corrections are
set to zero and thus dxq and dyq are not included in the adjustment. Note also
in this table that the elements which have been left blank are zeros.
Figure 16.3 Horizontal network.
16.6
ADJUSTMENT OF NETWORKS
293
TABLE 16.2 Data for Example 16.2
Distance observations
Occupied, i
Sighted, j
Distanceij (ft)
S (ft)
Q
R
S
T
Q
R
R
S
T
Q
S
T
1640.016
1320.001
1579.123
1664.524
2105.962
2266.035
0.026
0.024
0.025
0.026
0.029
0.030
Angle observations
Backsight, b
Instrument, i
Foresight, ƒ
Angle
S (ⴖ)
R
S
T
Q
R
S
S
S
Q
T
Q
R
Q
Q
Q
R
S
T
R
R
S
S
T
T
S
T
R
S
T
Q
T
Q
R
Q
R
S
38⬚48⬘50.7ⴖ
47⬚46⬘12.4ⴖ
273⬚24⬘56.5ⴖ
269⬚57⬘33.4ⴖ
257⬚32⬘56.8ⴖ
279⬚04⬘31.2ⴖ
42⬚52⬘51.0ⴖ
90⬚02⬘26.7ⴖ
51⬚08⬘45.0ⴖ
51⬚18⬘16.2ⴖ
46⬚15⬘02.0ⴖ
34⬚40⬘05.7ⴖ
4.0
4.0
4.4
4.7
4.7
4.5
4.3
4.5
4.3
4.0
4.0
4.0
Azimuth observations
From, i
To, j
Azimuth
S (ⴖ)
Q
R
0⬚06⬘24.5ⴖ
0.001
To fix the orientation of the network, the direction of course QR is included
as an observation, but with a very small estimated error, 0.001ⴖ. The last
row of Table 16.3 shows the inclusion of this constrained observation using
prototype equation (16.3). Since for azimuth QR only the foresight station,
R, is an unknown, only coefficients for the foresight station j are included in
the coefficient matrix.
Below are the necessary matrices for the first iteration when doing the
weighted least squares solution of the problem. Note that the numbers have
been truncated to five decimal places for publication purposes only. Following
these initial matrices, the results of the adjustment are listed, as determined
with program ADJUST.
294
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
TABLE 16.3 Format for Coefficient Matrix J of Example 16.4
Unknown
Observation
QR
RS
ST
TQ
QS
RT
∠RQS
∠SQT
∠TQR
∠QRS
∠RST
∠STQ
∠SRT
∠SRQ
∠QSR
∠TSQ
∠QTR
∠RTS
Az QR
0.00187
⫺1.00000
J⫽
0.00000
0.00000
0.00000
⫺0.73197
⫺125.77078
0.00000
125.77078
⫺125.58848
⫺0.18230
0.00000
61.83602
125.58848
0.18230
0.00000
62.01833
⫺62.01833
125.770798
dxr
dyr
j
i
j
i
i
b
i
b
ƒ
i
b
ƒ
i
b
i
i
ƒ
i
i
ƒ
ƒ
b
j
ƒ
b
j
dxs
dys
j
i
j
i
j
j
ƒ
b
ƒ
b
ƒ
i
b
b
b
i
i
ƒ
i
b
b
b
i
i
ƒ
ƒ
dxt
dyt
j
i
j
i
j
j
ƒ
b
ƒ
b
ƒ
i
ƒ
ƒ
i
ƒ
b
i
i
b
i
i
0.00000
0.00000
0.00000
0.00000
1.00000
0.00000
0.00000
⫺0.00117
1.00000
0.00117
⫺0.97673
0.21447
0.97673
⫺0.21447
0.00000
0.05772
0.99833
0.00000
0.00000
0.00000
0.00000
0.00000
0.77801
0.62825
0.00000
⫺0.68133
0.73197
0.00000
0.00000
0.68133
0.00000
0.00000
76.20105 ⫺61.53298
0.23544
61.53298
7.15291 ⫺123.71223
0.00000 ⫺76.20105
123.71223
⫺7.15291
0.00000
0.00000
⫺0.23544
0.00000
0.00000
⫺0.18230 ⫺156.26100
156.49644
184.27463 ⫺127.58038 ⫺28.01362
127.76269
⫺156.26100
134.73329 ⫺95.69861
0.00000 ⫺127.58038 ⫺28.01362
156.26100 ⫺62.01833 ⫺66.62776
0.18230
⫺89.63324
0.00000
0.00000
156.26100
0.18230
⫺156.49644
0.00000
0.00000
156.26100 ⫺76.38335 ⫺94.72803
28.01362
127.58038
0.00000 ⫺51.37934 ⫺89.54660
0.00000 ⫺69.17123
57.08446
0.00000
66.62776
28.01362 ⫺65.56206
38.61414
127.58038
⫺66.62776
0.00000
0.00000
0.00000
0.00000
⫺0.23544
16.6
ADJUSTMENT OF NETWORKS
295
The weight matrix is
1
0.0262
(zeros)
1
0.0242
1
0.0252
1
0.0262
1
0.0292
1
0.0302
1
4.02
W⫽
1
4.02
1
4.42
1
4.72
1
4.72
1
4.52
1
4.32
1
4.52
1
4.32
1
4.02
1
4.02
1
4.02
1
4.02
(zeros)
1
0.0012
296
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
The K matrix is
0.0031
⫺0.0099
⫺0.0229
⫺0.0007
⫺0.0053
⫺0.0196
⫺0.0090
⫺0.5988
0.2077
K⫽
⫺2.3832
1.4834
⫺1.4080
⫺1.0668
2.4832
⫺0.0742
⫺3.4092
⫺22.1423
2.4502
⫺0.3572
0.0000
Following is a summary of the results from ADJUST.
Number
Number
Number
Number
Number
of
of
of
of
of
Control Stations
Unknown Stations
Distance observations
Angle observations
Azimuth observations
»
»
»
»
»
1
3
6
12
1
*******************************************
Initial approximations for unknown stations
*******************************************
Station
X
Y
===========================
R 1,003.06 2,640.01
S 2,323.07 2,638.47
T 2,661.75 1,096.07
16.6
ADJUSTMENT OF NETWORKS
297
Control Stations
Station
X
Y
===========================
Q 1,000.00 1,000.00
*********************
Distance Observations
*********************
Station Station
Occupied Sighted
Distance
S
===================================
Q
R 1,640.016 0.026
R
S 1,320.001 0.024
S
T 1,579.123 0.025
T
Q 1,664.524 0.026
Q
S 2,105.962 0.029
R
T 2,266.035 0.030
******************
Angle Observations
******************
Station
Station
Station
Backsighted Occupied Foresighted
Angle
S
======================================================
R
Q
S
38⬚ 48⬘ 50.7ⴖ 4.0ⴖ
S
Q
T
47⬚ 46⬘ 12.4ⴖ 4.0ⴖ
T
Q
R 273⬚ 24⬘ 56.5ⴖ 4.4ⴖ
Q
R
S 269⬚ 57⬘ 33.4ⴖ 4.7ⴖ
R
S
T 257⬚ 32⬘ 56.8ⴖ 4.7ⴖ
S
T
Q 279⬚ 04⬘ 31.2ⴖ 4.5ⴖ
S
R
T
42⬚ 52⬘ 51.0ⴖ 4.3ⴖ
S
R
Q
90⬚ 02⬘ 26.7ⴖ 4.5ⴖ
Q
S
R
51⬚ 08⬘ 45.0ⴖ 4.3ⴖ
T
S
Q
51⬚ 18⬘ 16.2ⴖ 4.0ⴖ
Q
T
R
46⬚ 15⬘ 02.0ⴖ 4.0ⴖ
R
T
S
34⬚ 40⬘ 05.7ⴖ 4.0ⴖ
********************
Azimuth Observations
********************
Station Station
Occupied Sighted
Azimuth
S
===================================
Q
R
0⬚ 06⬘ 24.5 0.0ⴖ
298
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
Iteration 1
K MATRIX
0.0031
⫺0.0099
⫺0.0229
⫺0.0007
⫺0.0053
⫺0.0196
⫺0.0090
⫺0.5988
0.2077
⫺2.3832
1.4834
⫺1.4080
⫺1.0668
2.4832
⫺0.0742
⫺3.4092
⫺22.1423
2.4502
⫺0.3572
X MATRIX
⫺0.002906
⫺0.035262
⫺0.021858
0.004793
0.003996
⫺0.014381
Iteration 2
K MATRIX
0.0384
⫺0.0176
⫺0.0155
⫺0.0039
0.0087
⫺0.0104
⫺2.0763
⫺0.6962
2.3725
⫺0.6444
⫺0.5048
⫺3.3233
⫺1.3435
0.7444
3.7319
⫺5.2271
⫺18.5154
0.7387
0.0000
X MATRIX
0.000000
⫺0.000000
⫺0.000000
0.000000
⫺0.000000
0.000000
16.6
ADJUSTMENT OF NETWORKS
299
INVERSE MATRIX
0.00000000
0.00000047
0.00000003
0.00000034
0.00000005
0.00000019
0.00000047 0.00000003 0.00000034
0.00025290 0.00001780 0.00018378
0.00001780 0.00023696 ⫺0.00004687
0.00018378 ⫺0.00004687 0.00032490
0.00002767 0.00006675 0.00010511
0.00010155 ⫺0.00008552 0.00022492
0.00000005 0.00000019
0.00002767 0.00010155
0.00006675 ⫺0.00008552
0.00010511 0.00022492
0.00027128 0.00011190
0.00011190 0.00038959
*****************
Adjusted stations
*****************
Station
X
Y
Sx
Sy
=========================================
R 1,003.06 2,639.97 0.000 0.016
S 2,323.07 2,638.45 0.015 0.018
T 2,661.75 1,096.06 0.016 0.020
*******************************
Adjusted Distance Observations
*******************************
Station Station
Occupied Sighted
Distance
V
S
============================================
Q
R 1,639.978 ⫺0.0384 0.0159
R
S 1,320.019
0.0176 0.0154
S
T 1,579.138
0.0155 0.0158
T
Q 1,664.528
0.0039 0.0169
Q
S 2,105.953 ⫺0.0087 0.0156
R
T 2,266.045
0.0104 0.0163
***************************
Adjusted Angle Observations
***************************
Station Station
Station
Backsighted Occupied Foresighted
Angle
V
S
==========================================================
R
Q
S 38⬚ 48⬘ 52.8ⴖ 2.08ⴖ 1.75
S
Q
T 47⬚ 46⬘ 13.1ⴖ 0.70ⴖ 1.95
T
Q
R 273⬚ 24⬘ 54.1ⴖ ⫺2.37ⴖ 2.40
Q
R
S 269⬚ 57⬘ 34.0ⴖ 0.64ⴖ 2.26
R
S
T 257⬚ 32⬘ 57.3ⴖ 0.50ⴖ 2.50
S
T
Q 279⬚ 04⬘ 34.5ⴖ 3.32ⴖ 2.33
S
R
T 42⬚ 52⬘ 52.3ⴖ 1.34ⴖ 1.82
S
R
Q 90⬚ 02⬘ 26.0ⴖ ⫺0.74ⴖ 2.26
Q
S
R 51⬚ 08⬘ 41.3ⴖ ⫺3.73ⴖ 1.98
T
S
Q 51⬚ 18⬘ 21.4ⴖ 5.23ⴖ 2.04
300
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
Q
R
T
T
R
S
46⬚ 15⬘ 20.5ⴖ
34⬚ 40⬘ 05.0ⴖ
18.52 1.82
1.72
⫺0.74ⴖ
*****************************
Adjusted Azimuth Observations
*****************************
Station Station
Occupied Sighted
Azimuth
V
S
==========================================
Q
R 0⬚ 06⬘ 24.5ⴖ 0.00ⴖ 0.00ⴖ
***********************************
Adjustment Statistics
***********************************
Iterations ⫽ 2
Redundancies ⫽ 13
Reference Variance ⫽ 2.20
Reference So ⫽ 1.5
Passed X2 test at 99.0% significance level!
X2 lower value ⫽ 3.57
X2 upper value ⫽ 29.82
The a priori value of 1 used in computations involving
the reference variance.
Convergence!
16.7 2 TEST: GOODNESS OF FIT
At the completion of a least-squares adjustment, the significance of the computed reference variance, S 20, can be checked statistically. This check is often
referred to as a goodness-of-fit test since the computation of S 20 is based on
兺 v2. That is, as the residuals become larger, so will the reference variance
computed, and thus the model computed deviates more from the values observed. However, the size of the residuals is not the only contributing factor
to the size of the reference variance in a weighted adjustment. The stochastic
model also plays a role in the size of this value. Thus, when a 2 test indicates
that the null hypothesis should be rejected, it may be due to a blunder in the
data or an incorrect decision by the operator in selecting the stochastic model
for the adjustment. In Chapters 21 and 25 these matters are discussed in
greater detail. For now, the reference variance of the adjustment of Example
16.2 will be checked.
In Example 16.2 there are 13 degrees of freedom and the computed reference variance, S 20, is 2.2. In Chapter 10 it was shown that the a priori value
PROBLEMS
301
TABLE 16.4 Two-Tailed 2 Test on S 20
H0: S 2 ⫽ 1
Ha: S 2 ⫽ 1
Test statistic:
2 ⫽
S 2
13(2.2)
⫽
⫽ 28.6
2
1
Rejection region:
28.6 ⫽ 2 ⬎ 20.005,13 ⫽ 29.82
28.6 ⫽ 2 ⬍ 20.995,13 ⫽ 3.565
for the reference variance was 1. A check can now be made to compare the
computed value for the reference variance against its a priori value using a
two-tailed 2 test. For this adjustment, a significance level of 0.01 was selected. The procedures for doing the test were outlined in Section 5.4, and
the results for this example are shown in Table 16.4. Since ␣ /2 is 0.005 and
the adjustment had 13 redundant observations, the critical 2 value from the
table is 29.82. Now it can be seen that the 2 value computed is less that the
tabular value, and thus the test fails to reject the null hypothesis, H0. The
value of 1 for S 20 can and should be used when computing the standard deviations for the station coordinates and observations since the computed value
is only an estimate.
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
16.1
For the link traverse shown in Figure P16.1, assume that the distance
and angle standard deviations are 0.027 ft and 5ⴖ, respectively.
Using the control below, adjust the data given in the figure using
weighted least squares. The control station coordinates in units of feet
are
A: x ⫽ 944.79
Mk1: x ⫽ 991.31
y ⫽ 756.17
y ⫽ 667.65
C: x ⫽ 6125.48
Mk2: x ⫽ 6225.391
y ⫽ 1032.90
y ⫽ 1037.109
302
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of station B and give the standard
deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
Figure P16.1
16.2
Adjust by the method of least squares the closed traverse in Figure
P16.2. The data are given below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
Observed distances
Observed angles
Angle
Value
S (ⴖ)
Course
Distance (ft)
S (ft)
XAB
BAC
CBA
ACB
62⬚38⬘55.4ⴖ
56⬚18⬘41.9ⴖ
74⬚24⬘19.2ⴖ
49⬚16⬘55.9ⴖ
5.6
5.3
5.4
5.3
AB
BC
CA
1398.82
1535.70
1777.73
0.020
0.021
0.022
PROBLEMS
Control stations
303
Unknown stations
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
X
A
1490.18
1964.28
2063.39
1107.14
B
C
2791.96
3740.18
2234.82
1026.78
Figure P16.2
16.3
Adjust the network shown in Figure P16.3 by the method of least
squares. The data are listed below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
Unknown stations
Control station
Station
X (m)
Y (m)
Station
X (m
Y (m)
A
1776.596
2162.848
B
C
D
5339.61
5660.39
2211.95
2082.65
6103.93
6126.84
Distance observations
Angle observations
Course
Distance (m)
S (m)
Angle
Value
S (ⴖ)
AB
BC
CD
DA
AC
3563.905
4034.021
3448.534
3987.823
5533.150
0.013
0.014
0.013
0.014
0.018
DAC
CAB
ABC
BCA
ACD
DCA
38⬚18⬘44ⴖ
46⬚42⬘38ⴖ
93⬚16⬘18ⴖ
40⬚01⬘11ⴖ
45⬚47⬘57ⴖ
314⬚12⬘00ⴖ
4.0
4.0
4.0
4.0
4.0
4.0
304
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
The azimuth of line AB is 91⬚17⬘19.9ⴖ 0.001ⴖ.
Figure P16.3
16.4
Perform a weighted least squares adjustment using the data given in
Problem 15.4 and the additional distances given below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown station and the standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
Course
Distance (ft)
S (ft)
PU
QU
RU
SU
1214.44
1605.03
1629.19
1137.33
0.021
0.021
0.021
0.021
16.5
Do a weighted least squares adjustment using the data given in Problems 14.7 and 15.5.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown station and the standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
16.6
Using the program ADJUST, do a weighted least squares adjustment
using the data given in Problem 15.7 with the additional distances
given below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
PROBLEMS
16.7
16.8
Course
Distance (m)
S (m)
AD
AC
BD
BC
CD
17,337.708
27,331.345
23,193.186
19,382.380
17,745.364
0.087
0.137
0.116
0.097
0.089
305
Using the program ADJUST, do a weighted least squares adjustment
using the data given in Problem 15.9 with the additional distances
given below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
Course
Distance (m)
S (m)
AC
AD
BC
BD
CD
CE
CF
DE
DF
EF
EG
EH
FG
FH
9197.385
16,897.138
17,329.131
8666.341
14,384.926
8115.898
16,845.056
16,113.175
9019.629
13,800.459
8394.759
16,164.944
16,096.755
8170.129
0.028
0.051
0.052
0.026
0.043
0.025
0.051
0.049
0.027
0.042
0.026
0.049
0.048
0.025
Using the Program ADJUST, do a weighted least squares adjustment
using the data given in Problems 14.4 and 15.11.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown station and the standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
306
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
16.9
Using the program ADJUST, do a weighted least squares adjustment
using the data given below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
Unknown stations
Control station
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
A
108,250.29
33,692.06
B
C
D
E
F
G
H
104,352.50
106,951.03
155,543.53
160,220.88
154,763.88
131,436.82
129,558.23
54,913.38
75,528.38
75,701.62
57,165.44
57,165.44
54,645.29
61,487.41
Distance observations
Angle observations
Course
Distance (ft)
S (ft)
Angle
Value
S (ⴖ)
AB
BC
CD
DE
EF
FA
AG
GH
HD
21,576.31
20,778.13
48,592.81
19,117.21
5457.00
52,101.00
31,251.45
7095.33
29,618.90
0.066
0.063
0.146
0.059
0.020
0.157
0.094
0.024
0.090
BAG
ABC
BCD
HDC
DHG
HGA
EDH
DEF
EFA
GAF
58⬚18⬘15ⴖ
197⬚35⬘31ⴖ
262⬚36⬘41ⴖ
28⬚28⬘29ⴖ
103⬚19⬘34ⴖ
243⬚14⬘58ⴖ
75⬚28⬘59ⴖ
284⬚09⬘44ⴖ
153⬚13⬘19ⴖ
15⬚19⬘32ⴖ
3.0
3.0
3.0
3.0
3.0
3.0
3.0
3.0
3.0
3.0
16.10 Using the program ADJUST, do a weighted least squares adjustment
using the data given in Problem 15.12 and the distances listed below.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
(d) List the inverted normal matrix used in the last iteration.
PROBLEMS
307
Distance observations
Course
Distance (ft)
S (ft)
Course
Distance (ft)
S (ft)
AB
AE
EH
EI
EF
BF
DG
IF
FJ
JK
18,855.64
20,604.01
22,271.40
20,935.95
19,142.86
19,156.67
17,593.29
22,619.87
20,641.65
14,442.37
0.06
0.06
0.07
0.06
0.06
0.06
0.05
0.07
0.06
0.05
BE
BC
CF
FG
CD
CG
IJ
HI
JG
KG
19,923.80
17,548.84
18,009.19
15,957.20
15,359.20
18,077.08
17,962.96
17,984.70
21,035.70
19,528.95
0.06
0.05
0.06
0.05
0.05
0.06
0.06
0.06
0.06
0.06
16.11 Using the program ADJUST, do a weighted least squares adjustment
using the following.
(a) What is the reference standard deviation, S0?
(b) List the adjusted coordinates of the unknown stations and the
standard deviations.
(c) Tabulate the adjusted observations, the residuals, and the standard
deviations.
Unknown stations
Control stations
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
A
E
5,545.96
11,238.72
5504.56
7535.81
B
C
D
F
G
H
9949.16
5660.12
9343.18
8848.38
7368.43
6255.96
6031.81
8909.83
9642.46
6617.78
7154.46
6624.33
Distance observations
Course
Distance (ft)
S (ft)
Course
Distance (ft)
S (ft)
AB
BE
ED
DC
CA
FH
BD
4434.66
1981.15
2833.91
3989.03
3407.18
2592.35
3661.14
0.020
0.016
0.017
0.019
0.018
0.016
0.018
AH
HG
GC
GD
DF
FB
1325.89
1232.33
2623.49
3177.97
3064.98
1247.03
0.015
0.015
0.016
0.017
0.017
0.015
308
ADJUSTMENT OF HORIZONTAL SURVEYS: TRAVERSES AND NETWORKS
Angle observations
Stations
Angle
S (ⴖ)
Stations
Angle
S (ⴖ)
CAH
ABF
FBD
EDB
BDG
GCA
GHF
HGC
FGH
BFH
CGD
34⬚27⬘25ⴖ
34⬚51⬘20ⴖ
52⬚27⬘02ⴖ
32⬚27⬘10ⴖ
47⬚54⬘56ⴖ
45⬚50⬘58ⴖ
25⬚28⬘45ⴖ
67⬚24⬘11ⴖ
134⬚48⬘45ⴖ
152⬚07⬘08ⴖ
86⬚19⬘10ⴖ
3.4
3.5
3.5
3.1
3.1
3.2
3.5
3.5
3.7
3.5
3.2
HAB
ABE
DBE
BED
DCG
AHG
FHA
DGF
GFB
HFG
GDC
50⬚47⬘41ⴖ
137⬚26⬘38ⴖ
50⬚08⬘16ⴖ
97⬚24⬘33ⴖ
52⬚39⬘33ⴖ
212⬚17⬘20ⴖ
122⬚13⬘54ⴖ
71⬚27⬘44ⴖ
188⬚10⬘24ⴖ
19⬚42⬘31ⴖ
41⬚00⬘56ⴖ
3.4
3.2
3.2
3.3
3.1
3.8
3.5
3.3
3.7
3.4
3.1
For Problems 16.11 through 16.15, does the reference variance
computed for the adjustment pass the 2 test at a level of significance of 0.05?
16.12 Example 16.1
16.13 Problem 16.1
16.14 Problem 16.2
16.15 Problem 16.3
16.16 Problem 16.4
Programming Problems
16.17 Write a computational program that reads a file of station coordinates
and observations and then:
(a) writes the data to a file in a formatted fashion.
(b) computes the J, K, and W matrices.
(c) writes the matrices to a file that is compatible with the MATRIX
program.
(d) Demonstrate this program with Problem 16.6.
16.18 Write a program that reads a file containing the J, K, and W matrices
and then:
(a) writes these matrices in a formatted fashion.
(b) performs one iteration of Problem 16.6.
(c) writes the matrices used to compute the solution, and tabulates
the corrections to the station coordinates in a formatted fashion.
16.19 Write a program that reads a file of station coordinates and observations and then:
PROBLEMS
309
(a)
(b)
(c)
(d)
writes the data to a file in a formatted fashion.
computes the J, K, and W matrices.
performs a weighted least squares adjustment of Problem 16.6.
writes the matrices used in computations in a formatted fashion
to a file.
(e) computes the final adjusted station coordinates, their estimated
errors, the adjusted observations, their residuals, and their estimated errors, and writes them to a file in a formatted fashion.
16.20 Develop a computational program that creates the coefficient, weight,
and constant matrices for a network. Write the matrices to a file in a
format usable by the MATRIX program supplied with this book.
Demonstrate its use with Problem 16.6.
CHAPTER 17
ADJUSTMENT OF GPS NETWORKS
17.1
INTRODUCTION
For the past five decades, NASA and the U.S. military have been engaged in
a space research program to develop a precise positioning and navigation
system. The first-generation system, called TRANSIT, used six satellites and
was based on the Doppler principle. TRANSIT was made available for commercial use in 1967, and shortly thereafter its use in surveying began. The
establishment of a worldwide network of control stations was among its earliest and most valuable applications. Point positioning using TRANSIT required very lengthy observing sessions, and its accuracy was at the 1-m level.
Thus, in surveying it was suitable only for control work on networks consisting of widely spaced points. It was not satisfactory for everyday surveying
applications such as traversing or engineering layout.
Encouraged by the success of TRANSIT, a new research program was
developed that ultimately led to the creation of the NAVSTAR Global Positioning System (GPS). This second-generation positioning and navigation system utilizes a constellation of 24 orbiting satellites. The accuracy of GPS was
improved substantially over that of the TRANSIT system, and the disadvantage of lengthy observing sessions was also eliminated. Although developed
for military applications, civilians, including surveyors, also found uses for
the GPS system.
Since its introduction, GPS has been used extensively. It is reliable, efficient, and capable of yielding extremely high accuracies. GPS observations
can be taken day or night and in any weather conditions. A significant advantage of GPS is that visibility between surveyed points is not necessary.
Thus, the time-consuming process of clearing lines of sight is avoided. Al310
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
17.2
GPS OBSERVATIONS
311
though most of the earliest applications of GPS were in control work, improvements have now made the system convenient and practical for use in
virtually every type of survey, including property surveys, topographic mapping, and construction staking.
In this chapter we provide a brief introduction to GPS surveying. We explain the basic measurements involved in the system, discuss the errors in
those measurements, describe the nature of the adjustments needed to account
for those errors, and give the procedures for making adjustments of networks
surveyed using GPS. An example problem is given to demonstrate the procedures.
17.2
GPS OBSERVATIONS
Fundamentally, the global positioning system operates by observing distances
from receivers located on ground stations of unknown locations to orbiting
GPS satellites whose positions are known precisely. Thus, conceptually, GPS
surveying is similar to conventional resection, in which distances are observed
with an EDM instrument from an unknown station to several control points.
(The conventional resection procedure was discussed in Chapter 15 and illustrated in Example 15.2.) Of course, there are some differences between
GPS position determination and conventional resection. Among them is the
process of observing distances and the fact that the control stations used in
GPS work are satellites.
All of the 24 satellites in the GPS constellation orbit Earth at nominal
altitudes of 20,200 km. Each satellite continuously broadcasts unique electronic signals on two carrier frequencies. These carriers are modulated with
pseudorandom noise (PRN) codes. The PRN codes consist of unique sequences of binary values (zeros and ones) that are superimposed on the carriers. These codes appear to be random, but in fact they are generated
according to a known mathematical algorithm. The frequencies of the carriers
and PRN codes are controlled very precisely at known values.
Distances are determined in GPS surveying by taking observations on these
transmitted satellite signals. Two different observational procedures are used:
positioning by pseudoranging, and positioning by carrier-phase measurements. Pseudoranging involves determining distances (ranges) between satellites and receivers by observing precisely the time it takes transmitted
signals to travel from satellites to ground receivers. This is done by determining changes in the PRN codes that occur during the time it takes signals
to travel from the satellite transmitter to the antenna of the receiver. Then
from the known frequency of the PRN codes, very precise travel times are
determined. With the velocity and travel times of the signals known, the
pseudoranges can be computed. Finally, based on these ranges, the positions
of the ground stations can be calculated. Because pseudoranging is based on
312
ADJUSTMENT OF GPS NETWORKS
observing PRN codes, this GPS observation technique is also often referred
to as the code measurement procedure.
In the carrier-phase procedure, the quantities observed are phase changes
that occur as a result of the carrier wave traveling from the satellites to the
receivers. The principle is similar to the phase-shift method employed by
electronic distance-measuring instruments. However, a major difference is that
the satellites are moving, so that the signals cannot be returned to the transmitters for ‘‘true’’ phase-shift measurements. Instead, the phase shifts must
be observed at the receivers. But to make true phase-shift observations, the
clocks in the satellites and receivers would have to be perfectly synchronized,
which of course cannot be achieved. To overcome this timing problem and
to eliminate other errors in the system, differencing techniques (taking differences between phase observations) are used. Various differencing procedures can be applied. Single differencing is achieved by simultaneously
observing two satellites with one receiver. Single differencing eliminates satellite clock biases. Double differencing (subtracting the results of single differences from two receivers) eliminates receiver clock biases and other
systematic errors.
Another problem in making carrier-phase measurements is that only the
phase shift of the last cycle of the carrier wave is observed, and the number
of full cycles in the travel distance is unknown. (In EDM work this problem
is overcome by progressively transmitting longer wavelengths and observing
their phase shifts.) Again, because the satellites are moving, this cannot be
done in GPS work. However, by extending the differencing technique to what
is called triple differencing, this ambiguity in the number of cycles cancels
out of the solution. Triple differencing consists of differencing the results of
two double differences and thus involves making observations at two different
times to two satellites from two stations.
In practice, when surveys are done by observing carrier phases, four or
more satellites are observed simultaneously using two or more receivers located on ground stations. Also, the observations are repeated many times.
This produces a very large number of redundant observations, from which
many difference combinations can be computed.
Of the two GPS observing procedures, pseudoranging yields a somewhat
lower order of accuracy, but it is preferred for navigation use because it gives
instantaneous point positions of satisfactory accuracy. The carrier-phase technique produces a higher order of accuracy and is therefore the choice for
high-precision surveying applications. Adjustment of carrier-phase GPS observations is the subject of this chapter.
The differencing techniques used in carrier-phase observations, described
briefly above, do not yield positions directly for the points occupied by receivers. Rather, baselines (vector distances between stations) are determined.
These baselines are actually computed in terms of their coordinate difference
components ⌬X, ⌬Y, and ⌬Z. These coordinate differences are reported in the
17.2
GPS OBSERVATIONS
313
reference three-dimensional rectangular coordinate system described in Section 17.4.
To use the GPS carrier-phase procedure in surveying, at least two receivers
located on separate stations must be operated simultaneously. For example,
assume that two stations A and B were occupied for an observing session,
that station A is a control point, and that station B is a point of unknown
position. The session would yield coordinate differences ⌬XAB, ⌬YAB, and ⌬ZAB
between stations A and B. The X,Y,Z coordinates of station B can then be
obtained by adding the baseline components to the coordinates of A as
XB ⫽ XA ⫹ ⌬XAB
YB ⫽ YA ⫹ ⌬YAB
(17.1)
ZB ⫽ ZA ⫹ ⌬ZAB
Because carrier-phase observations do not yield point positions directly,
but rather, give baseline components, this method of GPS surveying is referred
to as relative positioning. In practice, often more than two receivers are used
simultaneously in relative positioning, which enables more than one baseline
to be determined during each observing session. Also, after the first observing
session, additional points are interconnected in the survey by moving the
receivers to nearby stations. In this procedure, at least one receiver is left on
one of the previously occupied stations. By employing this technique, a network of interconnected points can be created. Figure 17.1 illustrates an example of a GPS network. In this figure, stations A and B are control stations,
Figure 17.1 GPS survey network.
314
ADJUSTMENT OF GPS NETWORKS
and stations C, D, E, and F are points of unknown position. Creation of such
networks is a common procedure employed in GPS relative positioning work.
17.3
GPS ERRORS AND THE NEED FOR ADJUSTMENT
As in all types of surveying observations, GPS observations contain errors.
The principal sources of these errors are (1) orbital errors in the satellite, (2)
signal transmission timing errors due to atmospheric conditions, (3) receiver
errors, (4) multipath errors (signals being reflected so that they travel indirect
routes from satellite to receiver), and (5) miscentering errors of the receiver
antenna over the ground station and receiver height-measuring errors. To
account for these and other errors, and to increase the precisions of point
position, GPS observations are very carefully made according to strict
specifications, and redundant observations are taken. The fact that errors are
present in the observations makes it necessary to analyze the measurements
for acceptance or rejection. Also, because redundant observations have been
made, they must be adjusted so that all observed values are consistent.
In GPS surveying work where the observations are made using carrierphase observations, there are two stages where least squares adjustment is
applied. The first is in processing the redundant observations to obtain the
adjusted baseline components (⌬X, ⌬Y, ⌬Z), and the second is in adjusting
networks of stations wherein the baseline components have been observed.
The latter adjustment is discussed in more detail later in the chapter.
17.4 REFERENCE COORDINATE SYSTEMS FOR
GPS OBSERVATIONS
In GPS surveying, three different reference coordinate systems are involved.
First, the satellite positions at the instants of their observation are given in a
space-related Xs,Ys,Zs three-dimensional rectangular coordinate system. This
coordinate system is illustrated in Figure 17.2. In the figure, the elliptical
orbit of a satellite is shown. It has one of its two foci at G, Earth’s center of
gravity. Two points, perigee (point where the satellite is closest to G) and
apogee (point where the satellite is farthest from G), define the line of apsides.
This line, which also passes through the two foci, is the Xs axis of the satellite
reference coordinate system. The origin of the system is at G, the Ys axis is
in the mean orbital plane, and Zs is perpendicular to the Xs –Ys plane. Because
a satellite varies only slightly from its mean orbital plane, values of Zs are
small. For each specific instant of time that a given satellite is observed, its
coordinates are calculated in its unique Xs,Ys,Zs system.
In processing GPS observations, all Xs, Ys, and Zs coordinates that were
computed for satellite observations are converted to a common Earth-related
Xe,Ye,Ze three-dimensional geocentric coordinate system. This Earth-centered,
Earth-fixed coordinate system, illustrated in Figure 17.3, is also commonly
17.4
REFERENCE COORDINATE SYSTEMS FOR GPS OBSERVATIONS
315
Figure 17.2 Satellite reference coordinate system.
Figure 17.3 Earth-related three-dimensional coordinate system used in GPS carrierphase differencing computations.
316
ADJUSTMENT OF GPS NETWORKS
called the terrestrial geocentric system, or simply the geocentric system. It is
in this system that the baseline components are computed based on the differencing of observed carrier phase measurements. The origin of this coordinate system is at Earth’s gravitational center. The Ze axis coincides with
Earth’s Conventional Terrestrial Pole (CTP) axis, the Xe –Ye plane is perpendicular to the Ze axis, and the Xe axis passes through the Greenwich Meridian.
To convert coordinates from the space-related (Xs,Ys,Zs) system to the Earthcentered, Earth-related (Xe,Ye,Ze) geocentric system, six parameters are
needed. These are (a) the inclination angle i (the angle between the orbital
plane and Earth’s equatorial plane); (b) the argument of perigee (the angle
observed in the orbital plane between the equator and the line of apsides);
(c) right ascension of the ascending node (the angle observed in the plane
of Earth’s equator from the vernal equinox to the line of intersection between
the orbital and equatorial planes); (d) the Greenwich hour angle of the vernal
equinox (the angle observed in the equatorial plane from the Greenwich
meridian to the vernal equinox); (e) the semimajor axis of the orbital ellipse,
a; and (f) the eccentricity, e, of the orbital ellipse. The first four parameters
are illustrated in Figure17.3. For any satellite at any instant of time, these
four parameters are available. Software provided by GPS equipment manufacturers computes the Xs, Ys, and Zs coordinates of satellites at their instants
of observation, and it also transforms these coordinates into the Xe,Ye,Ze geocentric coordinate system used for computing the baseline components.
For the results of the baseline computations to be useful to local surveyors,
the Xe, Ye, and Ze coordinates must be converted to geodetic coordinates of
latitude, longitude, and height. The geodetic coordinate system is illustrated
in Figure 17.4, where the parameters are symbolized by , , and h, respectively. Geodetic coordinates are referenced to the World Geodetic System of
1984, which employs the WGS 84 ellipsoid. The center of this ellipsoid is
oriented at Earth’s gravitational center, and for most practical purposes it is
the same as the GRS 80 ellipsoid used for NAD 83. From latitude and longitude, state plane coordinates (which are more convenient for use by local
surveyors) can be computed.
It is important to note that geodetic heights are not orthometric heights
(elevations referred to the geoid). To convert geodetic heights to orthometric
heights, the geoid heights (vertical distances between the ellipsoid and geoid)
must be subtracted from geodetic heights.
17.5 CONVERTING BETWEEN THE TERRESTRIAL AND GEODETIC
COORDINATE SYSTEMS
GPS networks must include at least one control point, but more are preferable.
The geodetic coordinates of these control points will normally be given from
a previous GPS survey. Prior to processing carrier-phase observations to obtain adjusted baselines for a network, the coordinates of the control stations
17.5
CONVERTING BETWEEN THE TERRESTRIAL AND GEODETIC COORDINATE SYSTEMS
317
Figure 17.4 Geocentric coordinates (with the Earth-related Xe,Ye,Ze geocentric coordinate system superimposed).
in the network must be converted from their geodetic values into the Earthcentered, Earth-related Xe,Ye,Ze geocentric system. The equations for making
these conversions are
X ⫽ (N ⫹ h) cos cos
(17.2)
Y ⫽ (N ⫹ h) cos sin
(17.3)
Z ⫽ [N(1 ⫺ e2) ⫹ h] sin
(17.4)
In the equations above, h is the geodetic height of the point, the geodetic
latitude of the point, and the geodetic longitude of the point. Also, e is
eccentricity for the ellipsoid, which is computed as
e2 ⫽ 2ƒ ⫺ ƒ2
(17.5a)
a2 ⫺ b2
a2
(17.5b)
or
e2 ⫽
318
ADJUSTMENT OF GPS NETWORKS
where ƒ is the flattening factor of the ellipsoid; a and b are the semimajor
and semiminor axes, respectively, of the ellipsoid1; and N is the normal to
the ellipsoid at the point, which is computed as
N⫽
a
(17.6)
1 ⫺ e2 sin2
Example 17.1 Control stations A and B of the GPS network of Figure 17.1
have the following NAD83 geodetic coordinates:
A ⫽ 431546.2890ⴖ
A ⫽ ⫺895942.1640ⴖ
hA ⫽ 1382.618 m
B ⫽ 432346.3626ⴖ
B ⫽ ⫺895400.7570ⴖ
hB ⫽ 1235.457 m
Compute their Xe, Ye, and Ze geocentric coordinates.
SOLUTION For station A: By Equation (17.5a),
e2 ⫽
冉
冊
2
1
⫺
298.257223563
298.257223563
2
⫽ 0.006694379990
By Equation (17.6),
N⫽
6,378,137
⫽ 6,388,188.252 m
兹1 ⫺ e sin2(431546.2890ⴖ)
2
By Equation (17.2),
XA ⫽ (6,388,188.252 ⫹ 1382.618) cos(431546.2890ⴖ) cos(⫺895942.1640ⴖ)
⫽ 402.3509 m
By Equation (17.3),
YA ⫽ (6,388,188.252 ⫹ 1382.618) cos(431546.2890ⴖ) sin(⫺895942.1640ⴖ)
⫽ ⫺4,652,995.3011 m
1
The WGS 84 ellipsoid is used, whose a, b, and ƒ values are 6,378,137.0 m, 6,356,752.3142 m,
and 1 / 298.257223563, respectively.
17.5
CONVERTING BETWEEN THE TERRESTRIAL AND GEODETIC COORDINATE SYSTEMS
319
By Equation (17.4),
ZA ⫽ [6,388,188.252(1 ⫺ e2) ⫹ 1382.618] sin(431546.2890ⴖ)
⫽ 4,349,760.7775 m
For station B: Following the same procedure as above, the geocentric coordinates for station B are
XB ⫽ 8086.0318 m
YB ⫽ ⫺4,642,712.8474 m
ZB ⫽ 4,360,439.0833 m
After completing the network adjustment, it is necessary to convert all Xe,
Ye, and Ye geocentric coordinates to their geodetic values for use by local
surveyors. This conversion process follows these steps (refer to Figure 17.4):
Step 1: Determine the longitude, , from
Y
X
(17.7)
D ⫽ X 2 ⫹ Y 2
(17.8)
⫽ tan⫺1
Step 2: Compute D from
Step 3: Calculate an approximate latitude value 0 from
0 ⫽ tan⫺1
Z
D(1 ⫺ e2)
(17.9)
Step 4: Compute an approximate ellipsoid normal value N0 from
N0 ⫽
a
1 ⫺ e2 sin2 0
(17.10)
Step 5: Calculate an improved value for latitude 0 from
0 ⫽ tan⫺1
Z ⫹ e2N0 sin 0
D
(17.11)
320
ADJUSTMENT OF GPS NETWORKS
Step 6: Use the value of 0 computed in step 5, and return to step 4. Iterate
steps 4 and 5 until there is negligible change in 0. Using the values from
the last iteration for N0 and 0, the value for h is now computed2 as
h⫽
D
⫺ N0
cos 0
(17.12)
Example 17.2 Assume that the final adjusted coordinates for station C of
the network of Figure 17.4 were
XC ⫽ 12,046.5808 m
YC ⫽ ⫺4,649,394.0826 m
ZC ⫽ 4,353,160.0634 m
Compute the NAD83 geodetic coordinates for station C.
SOLUTION By Equation (17.7),
⫽ tan⫺1
⫺4,649,394.0826
12,046.5808
⫽ ⫺895105.5691ⴖ
By Equation (17.8),
D ⫽ (12,046.5808)2 ⫹ (⫺4,649,394.0826)2 ⫽ 4,649,409.6889 m
Using Equation (17.9), the initial value for 0 is
0 ⫽ tan⫺1
4,353,160.0634
⫽ 431826.2228ⴖ
D(1 ⫺ e2)
The first iteration for N0 and 0 is
N0 ⫽
6,378,137.0
⫽ 6,388,204.8545 m
1 ⫺ e sin2(431826.22280ⴖ)
0 ⫽ tan⫺1
2
4,353,160.0634 ⫹ e2 (6,388,204.8545) sin(431826.22280ⴖ)
D
⫽ 431826.1035ⴖ
The next iteration produced the final values as
2
Equation (17.12) is numerically stable for values of less than 45. For values of greater than
45, use the equation h ⫽ (Z / sin 0) ⫺ N0(1 ⫺ e2).
17.6
APPLICATION OF LEAST SQUARES IN PROCESSING GPS DATA
N0 ⫽ 6,388,204.8421
321
0 ⫽ 431826.1030ⴖ
Using Equation (17.12), the elevation of station C is
h⫽
D
⫺6,388,204.8421 ⫽ 1103.101 m
cos(431826.1030ⴖ)
A computer program included with the software package ADJUST will
make these coordinate conversions, both from geodetic to geocentric and from
geocentric to geodetic. These computations are also demonstrated in a Mathcad worksheet on the CD that accompanies this book.
17.6 APPLICATION OF LEAST SQUARES IN PROCESSING
GPS DATA
Least squares adjustment is used at two different stages in processing GPS
carrier-phase measurements. First, it is applied in the adjustment that yields
baseline components between stations from the redundant carrier-phase observations. Recall that in this procedure, differencing techniques are employed
to compensate for errors in the system and to resolve the cycle ambiguities.
In the solution, observation equations are written that contain the differences
in coordinates between stations as parameters. The reference coordinate system for this adjustment is the Xe,Ye,Ze geocentric system. A highly redundant
system of equations is obtained because, as described earlier, a minimum of
four (and often more) satellites are tracked simultaneously using at least two
(and often more) receivers. Furthermore, many repeat observations are taken.
This system of equations is solved by least squares to obtain the most probable X, Y, and Z components of the baseline vectors. The development
of these observation equations is beyond the scope of this book, and thus
their solution by least squares is also not covered herein.3
Software furnished by manufacturers of GPS receivers will process observed phase changes to form the differencing observation equations, perform
the least squares adjustment, and output the adjusted baseline vector components. The software will also output the covariance matrix, which expresses
the correlation between the X, Y, and Z components of each baseline.
The software is proprietary and thus cannot be included herein.
The second stage where least squares is employed in processing GPS observations is in adjusting baseline vector components in networks. This adjustment is made after the least squares adjustment of the carrier-phase
3
Readers interested in studying these observation equations should consult GPS Theory and Practice (Hoffman-Wellenhof et al., 2001) or GPS Satellite Surveying (Leick, 2004) (see the
bibliography).
322
ADJUSTMENT OF GPS NETWORKS
observations is completed. It is also done in the Xe,Ye,Ze geocentric coordinate
system. In network adjustments, the goal is to make all X coordinates (and
all X-coordinate differences) consistent throughout the figure. The same objective applies for all Y and Z coordinates. As an example, consider the GPS
network shown in Figure 17.1. It consists of two control stations and four
stations whose coordinates are to be determined. A summary of the baseline
observations obtained from the least squares adjustment of carrier-phase measurements for this figure is given in Table 17.1. The covariance matrix elements that are listed in the table are used for weighting the observations.
These are discussed in Section 17.8 but for the moment can be ignored.
A network adjustment of Figure 17.1 should yield adjusted X coordinates
for the stations (and adjusted coordinate differences between stations) that are
all mutually consistent. Specifically for this network, the adjusted X coordinate of station C should be obtained by adding XAC to the X coordinate of
station A; and the same value should be obtained by adding XBC to the X
coordinate of station B, or by adding XDC to the X coordinate of station D,
and so on. Equivalent conditions should exist for the Y and Z coordinates.
Note that these conditions do not exist for the data of Table 17.1, which
contains the unadjusted baseline measurements. The procedure of adjusting
GPS networks is described in detail in Section 17.8 and an example is given.
17.7
NETWORK PREADJUSTMENT DATA ANALYSIS
Prior to adjusting GPS networks, a series of procedures should be followed
to analyze the data for internal consistency and to eliminate possible blunders.
No control points are needed for these analyses. Depending on the actual
observations taken and the network geometry, these procedures may consist
of analyzing (1) differences between fixed and observed baseline components,
(2) differences between repeated observations of the same baseline components, and (3) loop closures. After making these analyses, a minimally constrained adjustment is usually performed that will help isolate any blunders
that may have escaped the first set of analyses. Procedures for making these
analyses are described in the following subsections.
17.7.1
Analysis of Fixed Baseline Measurements
GPS job specifications often require that baseline observations be taken between fixed control stations. The benefit of making these observations is to
verify the accuracy of both the GPS observational system and the control
being held fixed. Obviously, the smaller the discrepancies between observed
and known baseline lengths, the better. If the discrepancies are too large to
be tolerated, the conditions causing them must be investigated before proceeding. Note that in the data of Table 17.1, one fixed baseline (between
control points A and B) was observed. Table 17.2 gives the data for comparing
TABLE 17.1 Baseline Data Observed for the Network of Figure 17.1
(1)
(2)
(3)
(4)
(5)
From
To
X
Y
Z
A
A
B
B
D
D
F
F
F
F
F
B
A
Aa
C
E
C
D
C
E
A
C
E
D
B
F
F
B
3,601.2165
3,634.0754
⫺6,681.2467
⫺394.5204
⫺6,286.7054
⫺6,253.8534
⫺4,596.1610
⫺994.9377
⫺962.0694
5,291.7785
5,686.2926
⫺5,686.3033
4,596.1553
10,282.4550
3,399.2550
3,173.6652
⫺7,279.0148
⫺907.9593
⫺6,371.0583
⫺6,596.6697
⫺4355.9062
⫺956.6246
⫺1,182.2305
5,414.4311
6,322.3917
⫺6,322.3807
4,355.9141
10,678.3008
a
11,644.2232
⫺5,321.7164
3,960.5442
⫺11,167.6076
15,128.1647
⫺1,837.7459
⫺1,116.4523
10,527.7852
⫺6,438.1364
⫺4,600.3787
6,567.2311
⫺6,567.2310
1,116.4577
7,683.6883
Fixed baseline used only for checking, but not included in adjustment.
(6)
(7)
(8)
(9)
(10)
(11)
⫺9.520E-6
⫺2.100E-6
⫺2.230E-6
⫺2.720E-6
⫺1.440E-6
⫺1.210E-6
⫺8.100E-7
⫺2.270E-6
⫺8.900E-7
⫺9.900E-7
⫺6.400E-7
⫺6.300E-7
⫺8.200E-7
⫺7.290E-6
9.827E-4
2.005E-4
2.252E-4
2.670E-4
1.308E-4
1.283E-4
7.616E-5
2.397E-4
8.826E-5
1.204E-4
6.048E-5
6.629E-5
9.376E-5
7.310E-4
Covariance Matrix Elements
9.884E-4
2.158E-4
2.305E-4
2.700E-4
1.461E-4
1.231E-4
7.475E-5
2.567E-4
9.442E-5
9.330E-5
6.643E-5
5.512E-5
6.619E-5
7.2397E-4
⫺9.580E-6
⫺2.100E-6
⫺2.230E-6
⫺2.750E-6
⫺1.430E-6
⫺1.190E-6
⫺7.900E-7
⫺2.250E-6
⫺9.200E-7
⫺9.900E-7
⫺6.500E-7
⫺6.300E-7
⫺8.000E-7
⫺7.280E-6
9.520E-6
2.160E-6
2.070E-6
2.850E-6
1.340E-6
1.220E-6
8.800E-7
2.400E-6
1.040E-6
9.000E-7
6.900E-7
6.100E-7
9.000E-7
7.520E-6
9.377E-4
1.919E-4
2.546E-4
2.721E-4
1.614E-4
1.277E-4
6.593E-5
2.163E-4
9.959E-5
9.875E-5
7.465E-5
7.472E-5
8.108E-5
6.762E-4
323
324
ADJUSTMENT OF GPS NETWORKS
TABLE 17.2 Comparisons of Measured and Fixed Baseline Components
(1)
Component
(2)
Measured (m)
(3)
Fixed (m)
(4)
Difference (m)
(5)
ppma
X
Y
Z
7,683.6883
10,282.4550
10,678.3008
7,683.6809
10,282.4537
10,678.3058
0.0074
0.0013
0.0050
0.44
0.08
0.30
a
The total baseline length used in computing these ppm values was 16,697 m, which was derived
from the square root of the sum of the squares of X, Y, and Z values.
the observed and fixed baseline components. The observed values are listed
in column (2), and the fixed components are given in column (3). To compute
the fixed values, Xe, Ye, and Ze, geocentric coordinates of the two control
stations are first determined from their geodetic coordinates according to procedures discussed in Section 17.5. Then the X, Y, and Z differences between the Xe, Ye, and Ze coordinates for the two control stations are
determined. Differences (in meters) between the observed and fixed baseline
components are given in column (4). Finally, the differences, expressed in
parts per million (ppm), are listed in column (5). These ppm values are obtained by dividing column (4) differences by their corresponding total baseline
lengths and multiplying by 1,000,000.
17.7.2
Analysis of Repeat Baseline Measurements
Another procedure employed in evaluating the consistency of the data observed and in weeding out blunders is to make repeat observations of certain
baselines. These repeat observations are taken in different sessions and the
results compared. For example, in the data of Table 17.1, baselines AF and
BF were repeated. Table 17.3 gives comparisons of these observations using
the procedure that was used in Table 17.2. Again, the ppm values listed in
column (5) use the total baseline lengths in the denominator, which are com-
TABLE 17.3 Comparison of Repeat Baseline Measurements
Component
First
Observation
Second
Observation
Difference
(m)
ppm
XAF
YAF
ZAF
XBF
YBF
ZBF
1116.4577
4596.1553
4355.9141
⫺6567.2310
⫺5686.3033
⫺6322.3807
⫺1116.4523
⫺4596.1610
⫺4355.9062
0.0054
0.0057
0.0079
0.000l
0.0107
0.0110
0.84
0.88
1.23
0.01
1.00
1.02
6567.2311
5686.2926
6322.3917
17.7
NETWORK PREADJUSTMENT DATA ANALYSIS
325
puted from the square root of the sum of the squares of the measured baseline
components.
The Federal Geodetic Control Subcommittee (FGCS) has developed the
document Geometric Geodetic Accuracy Standards and Specifications for Using GPS Relative Positioning Techniques. It is intended to serve as a guideline
for planning, executing, and classifying geodetic surveys performed by GPS
relative positioning methods. This document may be consulted to determine
whether or not the ppm values of column (5) are acceptable for the required
order of accuracy for the survey. Besides ppm requirements, the FGCS guidelines specify other criteria that must be met for the various orders of accuracy
in connection with repeat baseline observations. It is wise to perform repeat
observations at the end of each day to check the repeatability of the software,
hardware, and field procedures.
17.7.3
Analysis of Loop Closures
GPS networks will typically consist of many interconnected closed loops. For
example, in the network of Figure 17.1, a closed loop is formed by points
ACBDEA. Similarly, ACFA, CFBC, BDFB, and so on, are other closed loops.
For each closed loop, the algebraic sum of the X components should equal
zero. The same condition should exist for the Y and Z components. These
loop misclosure conditions are very similar to the leveling loop misclosures
imposed in differential leveling and latitude and departure misclosures imposed in closed-polygon traverses. An unusually large misclosure within any
loop will indicate that either a blunder or a large random error exists in one
(or more) of the baselines of the loop.
To compute loop misclosures, the baseline components are simply added
algebraically for the loop chosen. For example, the misclosure in X components for loop ACBDEA would be computed as
cx ⫽ XAC ⫹ XCB ⫹ XBD ⫹ XDE ⫹ XEA
(17.13)
where cx is the loop misclosure in X coordinates. Similar equations apply for
computing misclosures in Y and Z coordinates.
Substituting numerical values into Equation (17.13), the misclosure in X
coordinates for loop ACBDEA is
cx ⫽ 11,644.2232 ⫺ 3960.5442 ⫺ 11,167.6076 ⫺ 1837.7459 ⫹ 5321.7164
⫽ 0.0419 m
Similarly, misclosures in Y and Z coordinates for that loop are
326
ADJUSTMENT OF GPS NETWORKS
cy ⫽ 3601.2165 ⫹ 6681.2467 ⫺ 394.5204 ⫺ 6253.8534 ⫺ 3634.0754
⫽ 0.0140 m
cz ⫽ 3399.2550 ⫹ 7279.0148 ⫺ 907.9593 ⫺ 6596.6697 ⫺ 3173.6652
⫽ ⫺0.0244 m
For evaluation purposes, loop misclosures are expressed in terms of the
ratios of resultant misclosures to the total loop lengths. They are given in
ppm. For any loop, the resultant misclosure is the square root of the sum of
the squares of its cx, cy, and cz values, and for loop ACBDEA the resultant
is 0.0505 m. The total length of a loop is computed by summing its legs,
each leg being computed from the square root of the sum of the squares of
its observed X, Y, and Z values. For loop ACBDEA, the total loop length
is 50,967 m, and the misclosure ppm ratio is therefore (0.0505/50,967) ⫻
1,000,000 ⫽ 0.99 ppm. Again, these ppm ratios can be compared against
values given in the FGCS guidelines to determine if they are acceptable for
the order of accuracy of the survey. As was the case with repeat baseline
observations, the FGCS guidelines also specify other criteria that must be met
in loop analyses besides the ppm values.
For any network, enough loop closures should be computed so that every
baseline is included within at least one loop. This should expose any large
blunders that exist. If a blunder does exist, its location can often be determined through additional loop-closure analyses. For example, assume that the
misclosure of loop ACDEA discloses the presence of a blunder. By also computing the misclosures of loops AFCA, CFDC, DFED, and EFAE, the baseline
containing the blunder can often be detected. In this example, if a large misclosure were found in loop DFED and all other loops appeared to be blunder
free, the blunder would be in line DE.
A computer program included within the software package ADJUST will
make these loop-closure computations. Documentation on the use of this program is given in its help file.
17.7.4
Minimally Constrained Adjustment
Prior to making the final adjustment of baseline observations in a network, a
minimally constrained least squares adjustment is usually performed. In this
adjustment, sometimes called a free adjustment, any station in the network
may be held fixed with arbitrary coordinates. All other stations in the network
are therefore free to adjust as necessary to accommodate the baseline observations and network geometry. The residuals that result from this adjustment
are strictly related to the baseline observations and not to faulty control coordinates. These residuals are examined and, from them, blunders that may
17.8
LEAST SQUARES ADJUSTMENT OF GPS NETWORKS
327
have gone undetected through the first set of analyses can be found and
eliminated.
17.8
LEAST SQUARES ADJUSTMENT OF GPS NETWORKS
As noted earlier, because GPS networks contain redundant observations, they
must be adjusted to make all coordinate differences consistent. In applying
least squares to the problem of adjusting baselines in GPS networks, observation equations are written that relate station coordinates to the coordinate
differences observed and their residual errors. To illustrate this procedure,
consider the example of Figure 16.1. For line AC of this figure, an observation
equation can be written for each baseline component observed as
XC ⫽ XA ⫹ XAC ⫹ vXAC
YC ⫽ YA ⫹ YAC ⫹ vYAC
(17.14)
ZC ⫽ ZA ⫹ ZAC ⫹ vZAC
Similarly, the observation equations for the baseline components of line CD
are
XD ⫽ XC ⫹ XCD ⫹ vXCD
YD ⫽ YC ⫹ YCD ⫹ vYCD
(17.15)
ZD ⫽ ZC ⫹ ZCD ⫹ vZCD
Observation equations of the foregoing form would be written for all measured baselines in any figure. For Figure 17.1, a total of 13 baselines were
observed, so the number of observation equations that can be developed is
39. Also, stations C, D, E, and F each have three unknown coordinates, for
a total of 12 unknowns in the problem. Thus, there are 39 ⫺ 12 ⫽ 27 redundant observations in the network. The 39 observation equations can be expressed in matrix form as
AX ⫽ L ⫹ V
(17.16)
If the observation equations for adjusting the network of Figure 17.1 are
written in the order in which the observations are listed in Table 17.1, the A,
X, L, and V matrices would be
328
ADJUSTMENT OF GPS NETWORKS
1
0
0
0
0
0
1
0
0
0
0
0
1
0
0
A⫽
0
0
0
0
0
0
1
0
0
0
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
1 0 0 0 0 0 0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
0 1 0 0 0 0 0
0 0 1 0 0 0 0
0 0 0 1 0 0 0
0 ⫺1 0 0 0 0 0
0 0 ⫺1 0 0 0 0
0 0 0 ⫺1 0 0 0
0 ⫺1 0 0 0 0 0
0 0 ⫺1 0 0 0 0
1 0 0 ⫺1 0 0 0
0 0 0 0 ⫺1 0 0
0 0 0 0 0 ⫺1 0
0 0 0 0 0 0 ⫺1
0 0 0 0 ⫺1 0 0
0 0 0 0 0 ⫺1 0
0 0 0 0 0 0 ⫺1
0 0 0 0 ⫺1 0 0
0 0 0 0 0 ⫺1 0
1 0 0 0 0 0 ⫺1
12046.5741
⫺4649394.0846
XC
YC
ZC
XE
YE
Z
X⫽ E L⫽
XD
YD
ZD
XF
YF
ZF
⯗
000000
000000
000000
0
0
0
4353160.0325
⫺4919.3655
⫺4649361.2257
4352934.4427
12046.5760
⫺4649394.0941
4353160.0685
⫺3081.5758
⫺46443107.3678
4359531.1240
15128.1647
⫺6286.7054
⫺6371.0583
⫺1837.7459
⫺6253.8534
⫺6596.6697
⫺1518.8032
4648399.1401
⫺4354116.6737
10527.7852
⫺994.9377
⫺956.6246
⫺6438.1364
⫺962.0694
⫺1182.2305
⯗
0
0
0
0
0
0
1
0
0
0
1
0
0
0
1
1518.8086
⫺4648399.1458
4354116.6916
vXAC
vYAC
vZAC
vXAE
vYAE
vZAE
vXBc
vYBC
vZBC
vXBD
vYBD
vZBD
vXDC
vYDC
vZDC
V ⫽ vXDE
vYDE
vZDE
vXFA
vYFA
vZFA
vXFC
vYFC
vZFC
vXFE
vYFE
vZFE
⯗
vXAF
vYAF
vZAF
The numerical values of the elements of the L matrix are determined by
rearranging the observation equations. Its first three elements are for the X,
Y, and Z baseline components of line AC, respectively. Those elements are
calculated as follows:
LX ⫽ XA ⫹ XAC
LY ⫽ YA ⫹ YAC
(17.17)
LZ ⫽ ZA ⫹ ZAC
The other elements of the L matrix are formed in the same manner as
demonstrated for baseline AC. However, before numerical values for the Lmatrix elements can be obtained, the Xe, Ye, and Ze geocentric coordinates of
all control points in the network must be computed. This is done by following
the procedures described in Section 17.5 and demonstrated by Example 17.1.
That example problem provided the Xe, Ye, and Ze coordinates of control
points A and B of Figure 17.1, which are used to compute the elements of
the L matrix given above.
17.8
LEAST SQUARES ADJUSTMENT OF GPS NETWORKS
329
Note that the observation equations for GPS network adjustment are linear
and that the only nonzero elements of the A matrix are either 1 and ⫺1. This
is the same type of matrix that was developed in adjusting level nets by least
squares. In fact, GPS network adjustments are performed in the very same
manner as level net adjustments, with the exception of the weights. In GPS
relative positioning, the three observed baseline components are correlated.
Therefore, a covariance matrix of dimensions 3 ⫻ 3 is derived for each baseline as a product of the least squares adjustment of the carrier-phase measurements. This covariance matrix is used to weight the observations in the
network adjustment in accordance with Equation (10.4). The weight matrix
for any GPS network is therefore a block-diagonal type, with an individual 3
⫻ 3 matrix for each baseline observed on the diagonal. When more than two
receivers are used, additional 3 ⫻ 3 matrices are created in the off-diagonal
region of the matrix to provide the correlation that exists between baselines
observed simultaneously. All other elements of the matrix are zeros.
The covariances for the observations in Table 17.1 are given in columns
(6) through (11). Only the six upper-triangular elements of the 3 ⫻ 3 covariance matrix for each observation are listed. This gives complete weighting
information, however, because the covariance matrix is symmetrical. Columns
6 through 11 list 2x , xy, xz, 2y , yz, and 2z , respectively. Thus, the full 3 ⫻
3 covariance matrix for baseline AC is
⌺AC
冤
9.884 ⫻ 10⫺4
⫽ ⫺9.580 ⫻ 10⫺4
9.520 ⫻ 10⫺4
⫺9.580 ⫻ 10⫺6
9.377 ⫻ 10⫺6
⫺9.520 ⫻ 10⫺6
9.520 ⫻ 10⫺6
⫺9.520 ⫻ 10⫺6
9.827 ⫻ 10⫺6
冥
The complete weight matrix for the example network of Figure 17.1 has
dimensions of 39 ⫻ 39. After inverting the full matrix and multiplying by
the a priori estimate for the reference variance, S20, in accordance with Equation (10.4), the weight matrix for the network of Figure 17.1 is (note that
S20 is taken as 1.0 for this computation and that no correlation between baselines is included):
W⫽
1011.8
10.2 ⫺9.7
0
0
0
0
0
0
10.2 1066.6
10.2
0
0
0
0
0
0
⫺9.7
10.2 1017.7
0
0
0
0
0
0
0
0
0 4634.5
50.2 ⫺49.4
0
0
0
0
0
0
50.2 5209.7
54.0
0
0
0
0
0
0
⫺49.4
54.0 4988.1
0
0
0
0
0
0
0
0
0 4339.1
37.7 ⫺39.5
0
0
0
0
0
0
37.7 3927.8
38.5
0
0
0
0
0
0
⫺39.5
38.5 4441.0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
䡠䡠
䡠
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1511.8
147.7 ⫺143.8
147.7 12336.0
106.5
⫺143.8
106.5 10667.8
330
ADJUSTMENT OF GPS NETWORKS
The system of observation equations (17.8) is solved by least squares using
Equation (11.35). This yields the most probable values for the coordinates of
the unknown stations. The complete output for the example of Figure 17.1
obtained using the program ADJUST follows.
****************
Control stations
****************
Station
X
Y
Z
=================================================
A
402.35087 ⫺4652995.30109 4349760.77753
B
8086.03178 ⫺4642712.84739 4360439.08326
****************
Distance Vectors
****************
From To
X
Y
Z
Covariance matrix elements
=============================================================================================
A C 11644.2232 3601.2165 3399.2550 9.884E-4 ⫺9.580E-6 9.520E-6 9.377E-4 ⫺9.520E-6 9.827E-4
A E ⫺5321.7164 3634.0754 3173.6652 2.158E-4 ⫺2.100E-6 2.160E-6 1.919E-4 ⫺2.100E-6 2.005E-4
B C
3960.5442 ⫺6681.2467 ⫺7279.0148 2.305E-4 ⫺2.230E-6 2.070E-6 2.546E-4 ⫺2.230E-6 2.252E-4
B D ⫺11167.6076 ⫺394.5204 ⫺907.9593 2.700E-4 ⫺2.750E-6 2.850E-6 2.721E-4 ⫺2.720E-6 2.670E-4
D C 15128.1647 ⫺6286.7054 ⫺6371.0583 1.461E-4 ⫺1.430E-6 1.340E-6 1.614E-4 ⫺1.440E-6 1.308E-4
D E ⫺1837.7459 ⫺6253.8534 ⫺6596.6697 1.231E-4 ⫺1.190E-6 1.220E-6 1.277E-4 ⫺1.210E-6 1.283E-4
F A ⫺1116.4523 ⫺4596.1610 ⫺4355.8962 7.475E-5 ⫺7.900E-7 8.800E-7 6.593E-5 ⫺8.100E-7 7.616E-5
F C 10527.7852 ⫺994.9377 ⫺956.6246 2.567E-4 ⫺2.250E-6 2.400E-6 2.163E-4 ⫺2.270E-6 2.397E-4
F E ⫺6438.1364 ⫺962.0694 ⫺1182.2305 9.442E-5 ⫺9.200E-7 1.040E-6 9.959E-5 ⫺8.900E-7 8.826E-5
F D ⫺4600.3787 5291.7785 5414.4311 9.330E-5 ⫺9.900E-7 9.000E-7 9.875E-5 ⫺9.900E-7 1.204E-4
F B
6567.2311 5686.2926 6322.3917 6.643E-5 ⫺6.500E-7 6.900E-7 7.465E-5 ⫺6.400E-7 6.048E-5
B F ⫺6567.2310 ⫺5686.3033 ⫺6322.3807 5.512E-5 ⫺6.300E-7 6.100E-7 7.472E-5 ⫺6.300E-7 6.629E-5
A F
1116.4577 4596.1553 4355.9141 6.619E-5 ⫺8.000E-7 9.000E-7 8.108E-5 ⫺8.200E-7 9.376E-5
Normal Matrix
=============================================================================================
16093.0
148.0 ⫺157.3
0
0
0
⫺6845.9
⫺60.0
69.4 ⫺3896.2
⫺40.1
38.6
148.0 15811.5
159.7
0
0
0
⫺60.0 ⫺6195.0
⫺67.6
⫺40.1 ⫺4622.2
⫺43.4
⫺157.3
159.7 17273.4
0
0
0
69.4
⫺67.6 ⫺7643.2
38.6
⫺43.4 ⫺4171.5
0
0
0
23352.1
221.9
⫺249.8 ⫺8124.3
⫺75.0
76.5 ⫺10593.3
⫺96.8
123.8
0
0
0
221.9
23084.9
227.3
⫺75.0 ⫺7832.2
⫺73.2
⫺96.8 ⫺10043.0
⫺100.1
0
0
0
⫺249.8
227.3
24116.4
76.5
⫺73.2 ⫺7795.7
123.8
⫺100.1 ⫺11332.6
⫺6845.9
⫺60.0
69.4 ⫺8124.3
⫺75.0
76.5
29393.8
278.7 ⫺264.4 ⫺10720.0
⫺106.7
79.2
⫺60.0 ⫺6195.0
⫺67.6
⫺75.0 ⫺7832.2
⫺73.2
278.7
27831.6
260.2
⫺106.7 ⫺10128.5
⫺82.5
69.4
⫺67.6 ⫺7643.2
76.5
⫺73.2 ⫺7795.7
⫺264.4
260.2 27487.5
79.2
⫺82.5 ⫺8303.5
⫺3896.2
⫺40.1
38.6 ⫺10593.3
⫺96.8
123.8 ⫺10720.0
⫺106.7
79.2
86904.9
830.9
⫺874.3
⫺40.1 ⫺4622.2
⫺43.4
⫺96.8 ⫺10043.0
⫺100.1
⫺106.7 ⫺10128.5
⫺82.5
830.9
79084.9
758.1
38.6
⫺43.4 ⫺4171.5
123.8
⫺100.1 ⫺11332.6
79.2
⫺82.5 ⫺8303.5
⫺874.3
758.1
79234.9
Constant Matrix
========================================================
⫺227790228.2336
⫺23050461170.3104
23480815458.7631
⫺554038059.5699
⫺24047087640.5196
17.8
331
LEAST SQUARES ADJUSTMENT OF GPS NETWORKS
21397654262.6187
⫺491968929.7795
⫺16764436256.9406
16302821193.7660
⫺5314817963.4907
⫺250088821081.7488
238833986695.9468
X Matrix
=============
12046.5808
⫺4649394.0826
4353160.0634
⫺4919.3391
⫺4649361.2199
4352934.4534
⫺3081.5831
⫺4643107.3692
4359531.1220
1518.8012
⫺4648399.1453
4354116.6894
Degrees of Freedom ⫽ 27
Reference Variance ⫽ 0.6135
Reference So ⫽ Ⳳ0.78
***********************
Adjusted Distance Vectors
***********************
From To
X
Y
Z
Vx
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
A
C
11644.2232 3601.2165 3399.2550 0.00669
A
E
⫺5321.7164
3634.0754 3173.6652 0.02645
B
C
3960.5442 ⫺6681.2467 ⫺7279.0148 0.00478
B
D ⫺11167.6076 ⫺394.5204 ⫺907.9593 ⫺0.00731
D
C
15128.1647 ⫺6286.7054 ⫺6371.0583 ⫺0.00081
D
E
⫺1837.7459 ⫺6253.8534 ⫺6596.6697 ⫺0.01005
F
A
⫺1116.4523 ⫺4596.1610 ⫺4355.8962
0.00198
F
C
10527.7852 ⫺994.9377 ⫺956.6246 ⫺0.00563
F
E
⫺6438.1364 ⫺962.0694 ⫺1182.2305 ⫺0.00387
F
D
⫺4600.3787
5291.7785 5414.4311 ⫺0.00561
F
B
6567.2311 5686.2926 6322.3917 ⫺0.00051
B
F
⫺6567.2310 ⫺5686.3033 ⫺6322.3807
0.00041
A
F
1116.4577 4596.1553 4355.9141 ⫺0.00738
Vy
Vz
0.00203
0.00582
0.01153
⫺0.00136
⫺0.00801
0.00268
0.00524
0.00047
⫺0.00514
⫺0.00232
0.00534
0.00536
0.00046
0.03082
0.01068
⫺0.00511
⫺0.00194
⫺0.00037
0.00109
⫺0.01563
⫺0.00140
⫺0.00545
0.00156
0.00220
⫺0.01320
⫺0.00227
332
ADJUSTMENT OF GPS NETWORKS
***************************
Advanced Statistical Values
***************************
From
To
ⳲS
Slope Dist
Prec
========================================================
A
C
0.0116
12,653.537
1,089,000
A
E
0.0100
7,183.255
717,000
B
C
0.0116
10,644.669
916,000
B
D
0.0097
11,211.408
1,158,000
D
C
0.0118
17,577.670
1,484,000
D
E
0.0107
9,273.836
868,000
F
A
0.0053
6,430.014
1,214,000
F
C
0.0115
10,617.871
921,000
F
E
0.0095
6,616.111
696,000
F
D
0.0092
8,859.036
964,000
F
B
0.0053
10,744.076
2,029,000
B
F
0.0053
10,744.076
2,029,000
A
F
0.0053
6,430.014
1,214,000
********************
Adjusted Coordinates
********************
Station
X
Y
Z
Sx
Sy
Sz
================================================================================
A
402.35087 ⫺4,652,995.30109 4,349,760.77753
B
8,086.03178 ⫺4,642,712.84739 4,360,439.08326
C
12,046.58076 ⫺4,649,394.08256 4,353,160.06335 0.0067 0.0068 0.0066
E
⫺4,919.33908 ⫺4,649,361.21987 4,352,934.45341 0.0058 0.0058 0.0057
D
⫺3,081.58313 ⫺4,643,107.36915 4,359,531.12202 0.0055 0.0056 0.0057
F
1,518.80119 ⫺4,648,399.14533 4,354,116.68936 0.0030 0.0031 0.0031
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
17.1
Using the WGS 84 ellipsoid parameters, convert the following geodetic coordinates to geocentric coordinates for these points.
(a) latitude: 405916.2541ⴖ N
longitude: 755957.0024ⴖ W
height: 164.248 m
PROBLEMS
333
(b) latitude: 411553.0534ⴖ N
longitude: 900236.7203ⴖ W
height: 229.085 m
(c) latitude: 445745.3603ⴖ N
longitude: 661256.2437ⴖ W
height: 254.362 m
(d) latitude: 335806.8409ⴖ N
longitude: 1222742.0462ⴖ W
height: 364.248 m
17.2
Using the WGS 84 ellipsoid parameters, convert the following geocentric coordinates (in meters) to geodetic coordinates for these
points.
(a) X ⫽ ⫺426,125.836 Y ⫽ ⫺5,472,467.695 Z ⫽ 3,237,961.360
(b) X ⫽ ⫺2,623,877.827 Y ⫽ ⫺3,664,128.366 Z ⫽ 4,498,233.251
(c) X ⫽ ⫺11,190.917 Y ⫽ ⫺4,469,623.638 Z ⫽ 4,534,918.934
(d) X ⫽ 2,051,484.893 Y ⫽ ⫺5,188,627.173 Z ⫽ 3,080,194.963
17.3
Given the following GPS observations and geocentric control station
coordinates to Figure P17.3, what are the most probable coordinates
for stations B and C using a weighted least squares adjustment? (All
data were collected with only two receivers.)
Figure P17.3
Control stations
Station
DA
X (m)
Y (m)
Z (m)
1,177,425.88739
1,178,680.69374
⫺4,674,386.55849
⫺4,673,056.15318
4,162,989.78649
4,164,169.65655
The vector covariance matrices for the X, Y, and Z values (in
meters) given are as follows. For baseline AB:
X ⫽ ⫺825.5585
0.00002199
0.00000030
0.00000030
Y ⫽
492.7369
0.00002806 ⫺0.00000030
Z ⫽
788.9732
0.00003640
334
ADJUSTMENT OF GPS NETWORKS
For baseline BC:
X ⫽ 606.2113
0.00003096 ⫺0.00000029
0.00000040
Y ⫽ 558.8905
0.00002709 ⫺0.00000029
Z ⫽ 546.7241
0.00002591
For baseline CD:
X ⫽ 1474.1569
0.00004127 ⫺0.00000045
0.00000053
278.7786
0.00004315
⫺0.00000045
Y ⫽
Z ⫽ ⫺155.8336
0.00005811
For baseline AC:
X ⫽ ⫺219.3510
0.00002440 ⫺0.00000020
0.00000019
Y ⫽ 1051.6348
0.00001700 ⫺0.00000019
Z ⫽ 1335.6877
0.00002352
For baseline DB:
X ⫽ ⫺2080.3644
17.4
0.00003589 ⫺0.00000034
0.00000036
Y ⫽ ⫺837.6605
0.00002658 ⫺0.00000033
Z ⫽ ⫺390.9075
0.00002982
Given the following GPS observations and geocentric control station
coordinates to accompany Figure P17.4, what are the most probable
coordinates for stations B and C using a weighted least squares adjustment? (All data were collected with only two receivers.)
Figure P17.4
PROBLEMS
335
Control stations
Station
X (m)
Y (m)
Z (m)
A
D
593,898.8877
593,319.2704
⫺4,856,214.5456
⫺4,855,416.0310
4,078,710.7059
4,079,738.3059
The vector covariance matrices for the X, Y, and Z values (in
meters) given are as follows. For baseline AB:
X ⫽ 678.034
5.098E-6 ⫺1.400E-5
6.928E-6
Y ⫽ 1206.714
7.440E-5 ⫺3.445E-5
Z ⫽ 1325.735
2.018E-5
For baseline BC:
X ⫽ ⫺579.895
3.404-E6 2.057E-6 ⫺3.036E-7
Y ⫽
145.342
2.015E-5 ⫺1.147E-5
Z ⫽
254.820
1.873E-5
For baseline AC:
98.138
6.518E-6 ⫺1.163E-7 ⫺3.811E-7
Y ⫽ 1352.039
3.844E-5 ⫺1.297E-5
Z ⫽ 1580.564
2.925E-5
X ⫽
For baseline DC:
X ⫽ 677.758
9.347E-6 ⫺1.427E-5
8.776E-6
Y ⫽ 553.527
2.954E-5 ⫺1.853E-5
Z ⫽ 552.978
1.470E-5
For baseline DC:
X ⫽ 677.756
17.5
9.170E-6 ⫺1.415E-5
8.570E-6
Y ⫽ 553.533
3.010E-5 ⫺1.862E-5
Z ⫽ 552.975
1.460E-5
Given the following GPS observations and geocentric control station
coordinates to accompany Figure P17.5, what are the most probable
336
ADJUSTMENT OF GPS NETWORKS
coordinates for station E using a weighted least squares adjustment?
(All data were collected with only two receivers.)
Figure P17.5
Control stations
Station
X (m)
Y (m)
Z (m)
A
B
C
D
⫺1,683,429.825
⫺1,524,701.610
⫺1,480,308.035
⫺1,725,386.928
⫺4,369,532.522
⫺4,230,122.822
⫺4,472,815.181
⫺4,436,015.964
4,390,283.745
4,511,075.501
4,287,476.008
4,234,036.124
The vector covariance matrices for the X, Y, and Z values (in
meters) given are as follows. For baseline AE:
X ⫽ 94,208.555 0.00001287 ⫺0.00000016 ⫺0.00000019
Y ⫽ ⫺61,902.843
0.00001621 ⫺0.00000016
Z ⫽ ⫺24,740.272
0.00001538
For baseline BE:
X ⫽ ⫺64,519.667
0.00003017 ⫺0.00000026
0.00000021
Y ⫽ ⫺77,506.853
0.00002834 ⫺0.00000025
Z ⫽ ⫺96,051.488
0.00002561
For baseline CE:
X ⫽ ⫺108,913.237
0.00008656 ⫺0.00000081 ⫺0.00000087
Y ⫽
165,185.492
0.00007882 ⫺0.00000080
Z ⫽
127,548.005
0.00008647
For baseline DE:
PROBLEMS
337
X ⫽ 136,165.650
0.00005893 ⫺0.00000066 ⫺0.00000059
Y ⫽ 128,386.277
0.00006707 ⫺0.00000064
Z ⫽ 180,987.895
0.00005225
For baseline EA:
X ⫽ ⫺94,208.554
0.00002284
0.00000036 ⫺0.00000042
Y ⫽ ⫺61,902.851
0.00003826 ⫺0.00000035
Z ⫽ ⫺24,740.277
0.00003227
For baseline EB:
X ⫽ 64,519.650
0.00008244
0.00000081 ⫺0.00000077
Y ⫽ 77,506.866
0.00007737 ⫺0.00000081
Z ⫽ 96,051.486
0.00008483
For baseline EC:
X ⫽
108,913.236
0.00002784 ⫺0.00000036
0.00000038
Y ⫽ ⫺165,185.494
0.00003396 ⫺0.00000035
Z ⫽ ⫺127,547.991
0.00002621
For baseline ED:
X ⫽ ⫺136,165.658
Y ⫽ ⫺128,386.282
Z ⫽ ⫺180,987.888
17.6
0.00003024 ⫺0.00000037
0.00000031
0.00003940 ⫺0.00000036
0.00003904
Given the following GPS observations and geocentric control station
coordinates to accompany Figure P17.6, what are the most probable
coordinates for stations B, D, and E using a weighted least squares
adjustment? (All data were collected with only two receivers.)
Figure P17.6
338
ADJUSTMENT OF GPS NETWORKS
Control stations
Station
X (m)
Y (m)
Z (m)
A
C
⫺1,439,383.018
⫺1,454,936.177
⫺5,325,949.910
⫺5,240,453.494
3,190,645.563
3,321,529.500
The vector covariance matrices for the X, Y, and Z values (in
meters) given are as follows. For baseline AB:
X ⫽ ⫺118,616.114
8.145E-4 ⫺7.870E-6
7.810E-6
Y ⫽
71,775.010
7.685E-4 ⫺7.820E-6
Z ⫽
62,170.130
8.093E-4
For baseline BC:
X ⫽ 103,062.915
8.521E-4 ⫺8.410E-6
8.520E-6
Y ⫽ 13,721.432
8.040E-4 ⫺8.400E-6
Z ⫽ 68,713.770
8.214E-4
For baseline CD:
X ⫽ 106,488.952
7.998E-4 ⫺7.850E-6
7.560E-6
Y ⫽ ⫺41,961.364
8.443E-4 ⫺7.860E-6
Z ⫽ ⫺21,442.604
7.900E-4
For baseline DE:
X ⫽
⫺7.715
3.547E-4 ⫺3.600E-6
3.720E-6
Y ⫽ ⫺35,616.922
3.570E-4 ⫺3.570E-6
Z ⫽ ⫺57,297.941
3.512E-4
For baseline EA:
X ⫽ ⫺90,928.118
8.460E-4 ⫺8.380E-6
8.160E-6
Y ⫽ ⫺7,918.120
8.824E-4 ⫺8.420E-6
Z ⫽ ⫺52,143.439
8.088E-4
For baseline CE:
PROBLEMS
X ⫽ 106,481.283
17.7
7.341E-4 ⫺7.250E-6
339
7.320E-6
Y ⫽ ⫺77,578.306
7.453E-4 ⫺7.290E-6
Z ⫽ ⫺78,740.573
7.467E-4
Given the following GPS observations and geocentric control station
coordinates to accompany Figure P17.7, what are the most probable
coordinates for stations B, D, E, and F using a weighted least squares
adjustment? (All data were collected with only two receivers.)
Figure P17.7
Control stations
Station
X (m)
Y (m)
Z (m)
A
B
⫺1,612,062.639
⫺1,613,505.053
⫺4,384,804.866
⫺4,383,572.785
4,330,846.142
4,331,494.264
The vector covariance matrices for the X, Y, and Z values (in
meters) given are as follows. For baseline AB:
X ⫽ ⫺410.891
7.064E-5 ⫺6.500E-7
6.400E-7
Y ⫽
979.896
6.389E-5 ⫺6.400E-7
Z ⫽
915.452
6.209E-5
For baseline BC:
X ⫽ ⫺1031.538
1.287E-5 ⫺1.600E-7
1.900E-7
252.184
1.621E-5 ⫺1.600E-7
Z ⫽ ⫺267.337
1.538E-5
Y ⫽
For baseline CD:
X ⫽
23.227
1.220E-5 ⫺9.000E-8
7.000E-8
Y ⫽ ⫺1035.622
1.104E-5 ⫺9.000E-8
Z ⫽ ⫺722.122
9.370E-6
340
ADJUSTMENT OF GPS NETWORKS
For baseline DE:
X ⫽ 1039.772
5.335E-5 ⫺4.900E-7
5.400E-7
Y ⫽ ⫺178.623
4.731E-5 ⫺4.800E-7
⫺3.753
5.328E-5
Z ⫽
For baseline EF:
X ⫽ ⫺434.125
7.528E-5 ⫺8.300E-7
7.500E-7
Y ⫽
603.788
8.445E-5 ⫺8.100E-7
Z ⫽
566.518
6.771E-5
For baseline EB:
X ⫽ ⫺31.465
3.340E-5 ⫺4.900E-7
5.600E-7
Y ⫽ 962.058
5.163E-5 ⫺4.800E-7
Z ⫽ 993.212
4.463E-5
For baseline FA:
X ⫽ 1845.068
9.490E-6 ⫺9.000E-8
8.000E-8
Y ⫽ ⫺873.794
7.820E-6 ⫺9.000E-8
Z ⫽ ⫺221.422
1.031E-5
For baseline FB:
X ⫽ 402.650
1.073E-5 ⫺1.600E-7
1.800E-7
Y ⫽ 358.278
1.465E-5 ⫺1.600E-7
Z ⫽ 426.706
9.730E-6
For baseline FC:
X ⫽ ⫺628.888
5.624E-5 ⫺6.600E-7
5.700E-7
Y ⫽
610.467
6.850E-5 ⫺6.300E-7
Z ⫽
159.360
6.803E-5
For baseline FD:
PROBLEMS
X ⫽ ⫺605.648
8.914E-5 ⫺8.100E-7
341
8.200E-7
Y ⫽ ⫺425.139
8.164E-5 ⫺8.100E-7
Z ⫽ ⫺562.763
7.680E-5
17.8 Given the following GPS observations and geocentric control
station coordinates to accompany Figure P17.8, what are the most
probable coordinates for stations B, D, E, and F using a weighted
least squares adjustment? (All data were collected with only two
receivers.)
Figure P17.8
Control stations
Station
X (m)
Y (m)
Z (m)
A
C
⫺2,413,963.823
⫺2,413,073.302
⫺4,395,420.994
⫺4,393,796.994
3,930,059.456
3,932,699.132
The vector covariance matrices for the X, Y, and Z values (in
meters) given are as follows. For baseline AB:
X ⫽ 535.100
4.950E-6 ⫺9.000E-8
7.000E-8
Y ⫽ 974.318
7.690E-6 ⫺9.000E-8
Z ⫽ 1173.264
8.090E-6
For baseline BC:
X ⫽ 355.412
5.885E-5 ⫺6.300E-7
7.400E-7
Y ⫽ 649.680
7.168E-5 ⫺6.500E-7
Z ⫽ 1466.409
6.650E-5
For baseline CD:
342
ADJUSTMENT OF GPS NETWORKS
X ⫽ ⫺1368.545
6.640E-6 ⫺4.000E-8
7.000E-8
Y ⫽
854.284
4.310E-6 ⫺4.000E-8
Z ⫽
⫺71.080
2.740E-6
For baseline DE:
X ⫽ ⫺671.715
1.997E-5 ⫺2.500E-7
2.500E-7
Y ⫽ ⫺1220.263
2.171E-5 ⫺2.400E-7
Z ⫽ ⫺951.343
3.081E-5
For baseline EF:
X ⫽ ⫺374.515
4.876E-5 ⫺3.600E-7
3.400E-7
Y ⫽ ⫺679.553
2.710E-5 ⫺3.700E-7
Z ⫽ ⫺1439.338
3.806E-5
For baseline EA:
X ⫽
1149.724
8.840E-5 ⫺8.000E7
8.300E-7
Y ⫽ ⫺1258.018
7.925E-5 ⫺8.200E-7
Z ⫽ ⫺1617.250
6.486E-5
For baseline EB:
X ⫽ 1684.833
1.861E-5 ⫺1.600E-7
2.000E-7
Y ⫽ ⫺283.698
1.695E-5 ⫺1.600E-7
Z ⫽ ⫺443.990
1.048E-5
For baseline EC:
X ⫽ 2040.254
6.966E-5 ⫺6.400E-7
7.300E-7
Y ⫽ 365.991
5.665E-5 ⫺6.300E-7
Z ⫽ 1022.430
7.158E-5
For baseline FA:
X ⫽ 1524.252
2.948E-5 ⫺3.500E-7
3.300E-7
Y ⫽ ⫺578.473
3.380E-5 ⫺3.500E-7
Z ⫽ ⫺177.914
4.048E-5
PROBLEMS
343
Given the data in each problem and using the procedure discussed
in Section 17.7.2, analyze the repeated baselines.
17.9
Problem 17.4.
17.10 Problem 17.5.
Given the data in each problem and using the procedures discussed
in Section 17.7.3, analyze the closures of the loops.
17.11 Problem 17.3, loops ABCDA, ABCA, ACDA, and BCDB
17.12 Problem 17.4, loops ABCDA, ACBA, and ADCA
17.13 Problem 17.6, loops ABCDEA and ABCEA
17.14 Problem 17.7, loops ABEA, DEFD, BFCB, CDFC, and ABCDEA
17.15 Problem 17.8, loops ABFA, BFEB, and BCDEB
Use program ADJUST to do each problem.
17.16 Problem 17.6
17.17 Problem 17.7
17.18 Problem 17.8
17.19 Problem 17.14
17.20 Problem 17.15
Programming Problems
17.21 Write a computational package that reads a file of station coordinates
and GPS baselines and then
(a) writes the data to a file in a formatted fashion.
(b) computes the A, L, and W matrices.
(c) writes the matrices to a file that is compatible with the MATRIX
program.
(d) Demonstrate this program with Problem 17.8.
17.22 Write a computational package that reads a file containing the A, L,
and W matrices and then:
(a) writes these matrices in a formatted fashion.
(b) performs a weighted least squares adjustment.
(c) writes the matrices used to compute the solution and tabulates the
station coordinates in a formatted fashion.
(d) Demonstrate this program with Problem 17.8.
17.23 Write a computational package that reads a file of station coordinates
and GPS baselines and then:
344
ADJUSTMENT OF GPS NETWORKS
(a)
(b)
(c)
(d)
writes the data to a file in a formatted fashion.
computes the A, L, and W matrices.
performs a weighted least squares adjustment.
writes the matrices used in computations in a formatted fashion
to a file.
(e) computes the final station coordinates, their estimated errors, the
adjusted baseline vectors, their residuals, and their estimated errors, and writes them to a file in a formatted fashion.
(f) Demonstrate this program with Problem 17.8.
CHAPTER 18
COORDINATE TRANSFORMATIONS
18.1
INTRODUCTION
The transformation of points from one coordinate system to another is a
common problem encountered in surveying and mapping. For instance, a
surveyor who works initially in an assumed coordinate system on a project
may find it necessary to transfer the coordinates to the state plane coordinate
system. In GPS surveying and in the field of photogrammetry, coordinate
transformations are used extensively. Since the inception of the North American Datum of 1983 (NAD 83), many land surveyors, management agencies,
state departments of transportation, and others have been struggling with the
problem of converting their multitudes of stations defined in the 1927 datum
(NAD 27) to the 1983 datum. Although several mathematical models have
been developed to make these conversions, all involve some form of coordinate transformation. This chapter covers the introductory procedures of
using least squares to compute several well-known and often used transformations. More rigorous procedures, which employ the general least squares
procedure, are given in Chapter 22.
18.2 TWO-DIMENSIONAL CONFORMAL
COORDINATE TRANSFORMATION
The two-dimensional conformal coordinate transformation, also known as the
four-parameter similarity transformation, has the characteristic that true shape
is retained after transformation. It is typically used in surveying when converting separate surveys into a common reference coordinate system. This
transformation is a three-step process that involves:
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
345
346
COORDINATE TRANSFORMATIONS
1. Scaling to create equal dimensions in the two coordinate systems
2. Rotation to make the reference axes of the two systems parallel
3. Translations to create a common origin for the two coordinate systems
The scaling and rotation are each defined by one parameter. The translations involve two parameters. Thus, there are a total of four parameters in
this transformation. The transformation requires a minimum of two points,
called control points, that are common to both systems. With the minimum
of two points, the four parameters of the transformation can be determined
uniquely. If more than two control points are available, a least squares adjustment is possible. After determining the values of the transformation parameters, any points in the original system can be transformed.
18.3
EQUATION DEVELOPMENT
Figure 18.1(a) and (b) illustrate two independent coordinate systems. In these
systems, three common control points, A, B, and C, exist (i.e., their coordinates are known in both systems). Points 1 through 4 have their coordinates
known only in the xy system of Figure 18.1(b). The problem is to determine
their XY coordinates in the system of Figure 18.1(a). The necessary equations
are developed as follows.
Step 1: Scaling. To make line lengths as defined by the xy coordinate system
equal to their lengths in the XY system, it is necessary to multiply xy
coordinates by a scale factor, S. Thus, the scaled coordinates x⬘ and y⬘ are
Figure 18.1 Two-dimensional coordinate systems.
18.3
EQUATION DEVELOPMENT
x⬘ ⫽ S x
y⬘ ⫽ S y
347
(18.1)
Step 2: Rotation. In Figure 18.2, the XY coordinate system has been superimposed on the scaled x⬘y⬘ system. The rotation angle, , is shown between
the y⬘ and Y axes. To analyze the effects of this rotation, an X⬘Y⬘ system
was constructed parallel to the XY system such that its origin is common
with that of the x⬘y⬘ system. Expressions that give the (X⬘,Y⬘) rotated coordinates for any point (such as point 4 shown) in terms of its x⬘y⬘ coordinates are
X⬘ ⫽ x⬘ cos ⫺ y⬘ sin
Y⬘ ⫽ x⬘ sin ⫹ y⬘ cos
(18.2)
Step 3: Translation. To finally arrive at XY coordinates for a point, it is necessary to translate the origin of the X⬘Y⬘ system to the origin of the XY
system. Referring to Figure 18.2, it can be seen that this translation is
accomplished by adding translation factors as follows:
X ⫽ X⬘ ⫹ TX and Y ⫽ Y⬘ ⫹ TY
(18.3)
If Equations (18.1), (18.2), and (18.3) are combined, a single set of equations results that transform the points of Figure 18.1(b) directly into Figure
18.1(a) as
X ⫽ (S cos )x ⫺ (S sin )y ⫹ TX
Y ⫽ (S sin )x ⫹ (S cos )y ⫹ TY
Figure 18.2 Superimposed coordinate systems.
(18.4)
348
COORDINATE TRANSFORMATIONS
Now let S cos ⫽ a, S sin ⫽ b, TX ⫽ c, and TY ⫽ d and add residuals
to make redundant equations consistent. Then Equations (18.4) can be written
as
ax ⫺ by ⫹ c ⫽ X ⫹ vX
(18.5)
ay ⫹ bx ⫹ d ⫽ Y ⫹ vY
18.4
APPLICATION OF LEAST SQUARES
Equations (18.5) represent the basic observation equations for a twodimensional conformal coordinate transformation that have four unknowns:
a, b, c, and d. The four unknowns embody the transformation parameters S,
, Tx, and Ty. Since two equations can be written for every control point, only
two control points are needed for a unique solution. When more than two are
present, a redundant system exists for which a least squares solution can be
found. As an example, consider the equations that could be written for the
situation illustrated in Figure 18.1. There are three control points, A, B, and
C, and thus the following six equations can be written:
axa ⫺ bya ⫹ c ⫽ XA ⫹ vXA
aya ⫹ bxa ⫹ d ⫽ YA ⫹ vYA
axb ⫺ byb ⫹ c ⫽ XB ⫹ vXB
(18.6)
ayb ⫹ bxb ⫹ d ⫽ YB ⫹ vYB
axc ⫺ byc ⫹ c ⫽ XC ⫹ vXC
ayc ⫹ bxc ⫹ d ⫽ YC ⫹ vYC
Equations (18.6) can be expressed in matrix form as
(18.7)
AX ⫽ L ⫹ V
where
xa
ya
x
A⫽ b
yb
xc
yc
⫺ya
xa
⫺yb
xb
⫺yc
xc
1
0
1
0
1
0
0
1
0
1
0
1
a
b
X⫽
c
d
XA
YA
XB
L⫽
YB
XC
YC
V⫽
vXA
vYA
vXB
vYB
vXC
vYC
18.4
349
APPLICATION OF LEAST SQUARES
The redundant system is solved using Equation (11.32). Having obtained
the most probable values for the coefficients from the least squares solution,
the XY coordinates of any additional points whose coordinates are known in
the xy system can then be obtained by applying Equations (18.5) (where the
residuals are now considered to be zeros).
After the adjustment, the scale factor S and rotation angle can be computed with the following equations:
⫽ tan⫺1
S⫽
b
a
(18.8)
a
cos
Example 18.1 A survey conducted in an arbitrary xy coordinate system
produced station coordinates for A, B, and C as well as for stations 1 through
4. Stations A, B, and C also have known state plane coordinates, labeled E
and N. It is required to derive the state plane coordinates of stations 1 through
4. Table 18.1 is a tabulation of the arbitrary coordinates and state plane
coordinates.
SOLUTION A computer listing from program ADJUST is given below for
the problem. The output includes the input data, the coordinates of transformed points, the transformation coefficients, and their estimated errors. Note
that the program formed the A and L matrices in accordance with Equation
(18.7). After obtaining the solution using Equation (11.32), the program
solved Equation (18.8) to obtain the rotation angle and scale factor of the
transformation. A complete solution for this example is given in the Mathcad
worksheet on the CD that accompanies this book.
TABLE 18.1 Data for Example 18.1
Point
E
N
x
y
A
B
C
1
2
3
4
1,049,422.40
1,049,413.95
1,049,244.95
51,089.20
49,659.30
49,884.95
121.622
141.228
175.802
174.148
513.520
754.444
972.788
⫺128.066
187.718
135.728
⫺120.262
⫺192.130
⫺67.706
120.994
350
COORDINATE TRANSFORMATIONS
Two Dimensional Conformal Coordinate Transformation
-------------------------------------------------------ax ⫺ by ⫹ Tx ⫽ X ⫹ VX
bx ⫹ ay ⫹ Ty ⫽ Y ⫹ VY
A matrix
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
121.622
128.066
1.000
0.000
⫺128.066
121.622
0.000
1.000
141.228
⫺187.718
1.000
0.000
187.718
141.228
0.000
1.000
175.802
⫺135.728
1.000
0.000
135.728
175.802
0.000
1.000
L matrix
~~~~~~~~~~~
1049422.400
51089.200
1049413.950
49659.300
1049244.950
49884.950
Transformed Control Points
POINT
X
Y
VX
VY
-------------------------------------------------------A
1,049,422.400
51,089.200
⫺0.004
0.029
B
1,049,413.950
49,659.300
⫺0.101
0.077
C
1,049,244.950
49,884.950
0.105
⫺0.106
Transformation Parameters:
a⫽
⫺4.51249
b⫽
⫺0.25371
Tx ⫽ 1050003.715
Ty ⫽
50542.131
0.00058
0.00058
0.123
0.123
Rotation ⫽ 183⬚ 13⬘ 05.0ⴖ
Scale ⫽ 4.51962
Adjustment’s Reference Variance ⫽ 0.0195
Transformed Points
POINT
X
Y
x
y
-------------------------------------------------------1
1,049,187.361
51,040.629
0.135
0.135
2
1,047,637.713
51,278.829
0.271
0.271
3
1,046,582.113
50,656.241
0.368
0.368
4
1,045,644.713
49,749.336
0.484
0.484
18.5
TWO-DIMENSIONAL AFFINE COORDINATE TRANSFORMATION
The two-dimensional affine coordinate transformation, also known as the sixparameter transformation, is a slight variation from the two-dimensional con-
18.5
TWO-DIMENSIONAL AFFINE COORDINATE TRANSFORMATION
351
formal transformation. In the affine transformation there is the additional
allowance for two different scale factors; one in the x direction and the other
in the y direction. This transformation is commonly used in photogrammetry
for interior orientation. That is, it is used to transform photo coordinates from
an arbitrary measurement photo coordinate system to a camera fiducial system
and to account for differential shrinkages that occur in the x and y directions.
As in the conformal transformation, the affine transformation also applies two
translations of the origin, and a rotation about the origin, plus a small nonorthogonality correction between the x and y axes. This results in a total of
six unknowns. The mathematical model for the affine transformation is
ax ⫹ by ⫹ c ⫽ X ⫹ VX
dx ⫹ ey ⫹ ƒ ⫽ Y ⫹ VY
(18.9)
These equations are linear and can be solved uniquely when three control
points exist (i.e., points whose coordinates are known in the both systems).
This is because for each point, an equation set in the form of Equations (18.9)
can be written, and three points yield six equations involving six unknowns.
If more than three control points are available, a least squares solution can
be obtained. Assume, for example, that four common points (1, 2, 3, and 4)
exist. Then the equation system would be
ax1 ⫹ by1 ⫹ c ⫽ X1 ⫹ VX1
dx1 ⫹ ey1 ⫹ ƒ ⫽ Y1 ⫹ VY1
ax2 ⫹ by2 ⫹ c ⫽ X2 ⫹ VX2
dx2 ⫹ ey2 ⫹ ƒ ⫽ Y2 ⫹ VY2
ax3 ⫹ by3 ⫹ c ⫽ X3 ⫹ VX3
(18.10)
dx3 ⫹ ey3 ⫹ ƒ ⫽ Y3 ⫹ VY3
ax4 ⫹ dy4 ⫹ c ⫽ X4 ⫹ VX4
dx4 ⫹ ey4 ⫹ ƒ ⫽ Y4 ⫹ VY4
In matrix notation, Equations (18.10) are expressed as AX ⫽ L ⫹ V, where
x1
0
x2
0
x3
0
x4
0
y1
0
y2
0
y3
0
y4
0
1
0
1
0
1
0
1
0
0
x1
0
x2
0
x3
0
x4
0 0
y1 1
0 0
y2 1
0 0
y3 1
0 0
y4 1
X1
vX1
a
Y1
vY1
b
X2
vX2
c
Y2
vY2
⫽
⫹
d
X3
vX3
e
Y3
vY3
ƒ
X4
vX4
Y4
vY4
(18.11)
352
COORDINATE TRANSFORMATIONS
The most probable values for the unknown parameters are computed using
least squares equation (11.32). They are then used to transfer the remaining
points from the xy coordinate system to the XY coordinate system.
Example 18.2 Photo coordinates, which have been measured using a digitizer, must be transformed into the camera’s fiducial coordinate system. The
four fiducial points and the additional points were measured in the digitizer’s
xy coordinate system and are listed in Table 18.2 together with the known
camera XY fiducial coordinates.
SOLUTION A self-explanatory computer solution from the program ADJUST that yields the least squares solution for an affine transformation is
shown below. The complete solution for this example is given in the Mathcad
worksheet on the CD that accompanies this book.
Two Dimensional Affine Coordinate Transformation
-------------------------------------------------------ax ⫹ by ⫹ c ⫽ X ⫹ VX
dx ⫹ ey ⫹ f ⫽ Y ⫹ VY
A matrix
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
0.764
5.960
1.000
0.000
0.000
0.000
0.000
0.000
0.000
0.764
5.960
1.000
5.062
10.541
1.000
0.000
0.000
0.000
0.000
0.000
0.000
5.062
10.541
1.000
9.663
6.243
1.000
0.000
0.000
0.000
0.000
0.000
0.000
9.663
6.243
1.000
5.350
1.654
1.000
0.000
0.000
0.000
0.000
0.000
0.000
5.350
1.654
1.000
L matrix
~~~~~~~~
⫺113.000
⫺113.000
0.001
0.001
112.998
112.998
0.001
0.001
TABLE 18.2 Coordinates of Points for Example 18.2
1
3
5
7
306
307
X
Y
x
y
x
y
⫺113.000
0.003
112.993
0.003
⫺112.999
0.764
5.062
9.663
5.350
1.746
5.329
5.960
10.541
6.243
1.654
9.354
9.463
0.104
0.096
0.112
0.096
0.112
0.120
0.088
0.104
0.001
112.998
0.001
18.6
TWO-DIMENSIONAL PROJECTIVE COORDINATE TRANSFORMATION
353
Transformed Control Points
POINT
X
Y
VX
VY
-------------------------------------------------------1
⫺113.000
0.003
0.101
0.049
3
0.001
112.993
⫺0.086
⫺0.057
5
112.998
0.003
0.117
0.030
7
0.001
⫺112.999
⫺0.086
⫺0.043
Transformation Parameters:
a
b
c
d
e
f
⫽
⫽
⫽
⫽
⫽
⫽
25.37152
0.82220
⫺137.183
⫺0.80994
25.40166
⫺150.723
0.02532
0.02256
0.203
0.02335
0.02622
0.216
Adjustment’s Reference Variance ⫽ 2.1828
Transformed Points
POINT
X
Y
x
y
-------------------------------------------------------1
⫺112.899
0.052
0.132
0.141
3
⫺0.085
112.936
0.125
0.147
5
113.115
0.033
0.139
0.118
7
⫺0.085
⫺113.042
0.125
0.134
306
⫺85.193
85.470
0.134
0.154
307
5.803
85.337
0.107
0.123
18.6 TWO-DIMENSIONAL PROJECTIVE
COORDINATE TRANSFORMATION
The two-dimensional projective coordinate transformation is also known as
the eight-parameter transformation. It is appropriate to use when one twodimensional coordinate system is projected onto another nonparallel system.
This transformation is commonly used in photogrammetry and it can also be
used to transform NAD 27 coordinates into the NAD 83 system. In their final
form, the two-dimensional projective coordinate transformation equations are
354
COORDINATE TRANSFORMATIONS
X⫽
a1x ⫹ b1y ⫹ c
a3x ⫹ b3y ⫹ 1
(18.12)
a x ⫹ b2 y ⫹ c
Y⫽ 2
a3x ⫹ b3y ⫹ 1
Upon inspection, it can be seen that these equations are similar to the affine
transformation. In fact, if a3 and b3 were equal to zero, these equations are
the affine transformation. With eight unknowns, this transformation requires
a minimum of four control points (points having coordinates known in both
systems). If there are more than four control points, the least squares solution
can be used. Since these are nonlinear equations, they must be linearized and
solved using Equation (11.37) or (11.39). The linearized form of these equations is
⫽
⭸X
⭸a1
⭸X
⭸b1
0
0
0
0
⭸X
⭸c1
0
0
0
0
0
⭸X
⭸a3
⭸x
⭸a2
⭸X
⭸b2
⭸X
⭸c2
⭸X
⭸a3
0
0
0
0
⭸X
⭸b3
0
0
⭸X
⭸b3
0
X ⫺ X0
Y ⫺ Y0
da1
db1
dc1
da2
db2
dc2
da3
db3
(18.13)
where
⭸X
x
⭸X
y
⭸X
1
⫽
⫽
⫽
⭸a1
a3x ⫹ b3 ⫹ 1 ⭸b1 a3x ⫹ b3 ⫹ 1 ⭸c1 a3x ⫹ b3 ⫹ 1
⭸Y
x
⭸Y
y
⭸Y
1
⫽
⫽
⫽
⭸a2
a3x ⫹ b3 ⫹ 1 ⭸b2 a3x ⫹ b3 ⫹ 1 ⭸c2 a3x ⫹ b3 ⫹ 1
⭸X
a x ⫹ b1y ⫹ c1
⫽⫺ 1
x
⭸a3
(a3x ⫹ b3 ⫹ 1)2
⭸X
a x ⫹ b1y ⫹ c1
⫽⫺ 1
y
⭸b3
(a3x ⫹ b3 ⫹ 1)2
⭸Y
a x ⫹ b2 y ⫹ c2
⫽⫺ 2
x
⭸a3
(a3x ⫹ b3 ⫹ 1)2
⭸Y
a x ⫹ b2 y ⫹ c2
⫽⫺ 2
y
⭸b3
(a3x ⫹ b3 ⫹ 1)2
For each control point, a set of equations of the form of Equation (18.13)
can be written. A redundant system of equations can be solved by least
squares to yield the eight unknown parameters. With these values, the re-
18.6
TWO-DIMENSIONAL PROJECTIVE COORDINATE TRANSFORMATION
355
maining points in the xy coordinate system are transformed into the XY
system.
Example 18.3 Given the data in Table 18.3, determine the best-fit projective
transformation parameters and use them to transform the remaining points
into the XY coordinate system.
Program ADJUST was used to solve this problem and the results follow.
The complete solution for this example is given in the Mathcad worksheet on
the CD that accompanies this book.
Two Dimensional Projective Coordinate Transformation of
File
-------------------------------------------------------a1x ⫹ b1y ⫹ c1
---------------- ⫽ X ⫹ VX
a3x ⫹ b3y ⫹ 1
a2x ⫹ b2y ⫹ c2
---------------- ⫽ Y ⫹ VY
a3x ⫹ b3y ⫹ 1
Transformation Parameters:
a1
b1
c1
a2
b2
c2
a3
b3
⫽
⫽
⫽
⫽
⫽
⫽
⫽
⫽
25.00274
0.80064
⫺134.715
⫺8.00771
24.99811
⫺149.815
0.00400
0.00200
0.01538
0.01896
0.377
0.00954
0.01350
0.398
0.00001
0.00001
TABLE 18.3 Data for Example 18.3
Point
X
Y
x
y
x
y
1
2
3
4
5
6
7
8
1420.407
895.887
⫺944.926
968.084
1993.262
⫺3382.284
895.362
351.398
641.434
⫺1384.138
⫺2367.511
3487.762
90.0
50.0
⫺30.0
50.0
110.0
⫺100.0
⫺60.0
⫺100.0
90.0
40.0
20.0
⫺40.0
⫺80.0
80.0
20.0
⫺100.0
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
0.3
356
COORDINATE TRANSFORMATIONS
Adjustment’s Reference Variance ⫽ 3.8888
Number of Iterations ⫽ 2
Transformed Control Points
POINT
X
Y
VX
VY
-------------------------------------------------------1
1,420.165
895.444
⫺0.242
0.082
2
896.316
351.296
0.429
⫺0.102
3
⫺944.323
641.710
0.603
0.276
4
967.345
⫺1,384.079
⫺0.739
0.059
5
1,993.461
⫺2,367.676
0.199
⫺0.165
6
⫺3,382.534
3,487.612
⫺0.250
⫺0.150
Transformed Points
POINT
X
Y
x
y
-------------------------------------------------------1
1,420.165
895.444
0.511
0.549
2
896.316
351.296
0.465
0.458
3
⫺944.323
641.710
0.439
0.438
4
967.345
⫺1,384.079
0.360
0.388
5
1,993.461
⫺2,367.676
0.482
0.494
6
⫺3,382.534
3,487.612
0.558
0.563
7
⫺2,023.678
1,038.310
1.717
0.602
8
⫺6,794.740
⫺4,626.976
51.225
34.647
18.7 THREE-DIMENSIONAL CONFORMAL
COORDINATE TRANSFORMATION
The three-dimensional conformal coordinate transformation is also known as
the seven-parameter similarity transformation. It transfers points from one
three-dimensional coordinate system to another. It is applied in the process
of reducing data from GPS surveys and is also used extensively in photogrammetry. The three-dimensional conformal coordinate transformation involves seven parameters, three rotations, three translations, and one scale
factor. The rotation matrix is developed from three consecutive twodimensional rotations about the x, y, and z axes, respectively. Given in sequence, these are as follows.
In Figure 18.3, the rotation 1 about the x axis expressed in matrix form
is
18.7
THREE-DIMENSIONAL CONFORMAL COORDINATE TRANSFORMATION
357
Figure 18.3 1 rotation.
(a)
X1 ⫽ R1X0
where
1
R1 ⫽ 0
0
x1
X1 ⫽ y1
z1
0
0
cos 1 sin 1
⫺sin 1 cos 1
x
X0 ⫽ y
z
In Figure 18.4, the rotation 2 about the y axis expressed in matrix form is
(b)
X2 ⫽ R2X1
where
x2
X2 ⫽ y2
z2
and R2 ⫽
cos 2
0
sin 2
0
1
0
⫺sin 2
0
cos 2
In Figure 18.5, the rotation 3 about the z axis expressed in matrix form
is
X ⫽ R3X2
where
Figure 18.4 2 rotation.
(c)
358
COORDINATE TRANSFORMATIONS
Figure 18.5 3 rotation.
X⫽
X
Y
Z
and R3 ⫽
cos 3
⫺sin 3
0
sin 3 0
cos 3 0
0
1
Substituting Equation (a) into (b) and in turn into (c) yields
X ⫽ R3R2R1X0 ⫽ RX0
(d)
When multiplied together, the three matrices R3, R2, and R1 in Equation (d)
develop a single rotation matrix R for the transformation whose individual
elements are
r11
R ⫽ r21
r31
r12
r23
r32
r13
r23
r33
(18.14)
where
r11 ⫽ cos 2 cos 3
r12 ⫽ sin 1 sin 2 cos 3 ⫹ cos 1 sin 3
r13 ⫽ ⫺cos 1 sin 2 cos 3 ⫹ sin 1 sin 3
r21 ⫽ ⫺cos 2 sin 3
r22 ⫽ ⫺sin 1 sin 2 sin 3 ⫹ cos 1 cos 3
r23 ⫽ cos 1 sin 2 sin 3 ⫹ sin 1 cos 3
r31 ⫽ sin 2
r32 ⫽ ⫺sin 1 cos 2
r33 ⫽ cos 1 cos 2
Since the rotation matrix is orthogonal, it has the property that its inverse
is equal to its transpose. Using this property and multiplying the terms of the
matrix X by a scale factor, S, and adding translations factors Tx, Ty, and Tz to
18.7
359
THREE-DIMENSIONAL CONFORMAL COORDINATE TRANSFORMATION
translate to a common origin yields the following mathematical model for the
transformation:
X ⫽ S(r11x ⫹ r21y ⫹ r31z) ⫹ Tx
(18.15)
Y ⫽ S(r12x ⫹ r22 y ⫹ r32z) ⫹ Ty
Z ⫽ S(r13x ⫹ r23y ⫹ r33z) ⫹ Tz
Equations (18.15) involve seven unknowns (S, 1, 2, 3, Tx, Ty, Tz). For
a unique solution, seven equations must be written. This requires a minimum
of two control stations with known XY coordinates and also xy coordinates,
plus three stations with known Z and z coordinates. If there are more than
the minimum number of control points, a least-squares solution can be used.
Equations (18.15) are nonlinear in their unknowns and thus must be linearized
for a solution. The following linearized equations can be written for each
point as
⭸X
⭸S
⭸Y
⭸S
⭸Z
⭸S
⭸X
⭸2
0
0
0
⭸Y
⭸1
0
⭸Z
⭸1
0
⭸Y
⭸2
0
⭸Z
⭸2
0
⭸X
⭸3
0
0
⭸Y
⭸3
0
0
⭸Z
⭸3
0
1
0
0
0
1
0
0
0
1
dS
d1
d2
X ⫺ X0
d3 ⫽ Y ⫺ Y0
dTx
Z ⫺ Z0
dTy
dTz
(18.16)
where
⭸X
⫽ r11x ⫹ r21y ⫹ r31z
⭸S
⭸Y
⫽ r12x ⫹ r22 y ⫹ r32z
⭸S
⭸Y
⫽ ⫺S[r13x ⫹ r23y ⫹ r33z]
⭸1
⭸Z
⫽ r13x ⫹ r23y ⫹ r33z
⭸S
⭸Z
⫽ S[r12x ⫹ r22 y ⫹ r32z]
⭸1
⭸X
⫽ S(⫺x sin 2 cos 3 ⫹ y sin 2 sin 3 ⫹ z cos 2
⭸2
⭸Y
⫽ S(x sin 1 cos 2 cos 3 ⫺ y sin 1 cos 2 sin 3 ⫹ z sin 1 sin 2)
⭸2
⭸Z
⫽ S(⫺x cos 1 cos 2 cos 3 ⫹ y cos 1 cos 2 sin 3 ⫺ z cos 1 sin 2)
⭸2
⭸X
⫽ S(r21x ⫺ r11y)
⭸3
⭸Y
⫽ S(r22x ⫺r12 y)
⭸3
⭸Z
⫽ S(r23x ⫺ r13y)
⭸3
360
COORDINATE TRANSFORMATIONS
Example 18.4 The three-dimensional xyz coordinates were measured for six
points. Four of these points (1, 2, 3, and 4) were control points whose coordinates were also known in the XYZ control system. The data are shown in
Table 18.4. Compute the parameters of a three-dimensional conformal coordinate transformation and use them to transform points 5 and 6 in the XYZ
system.
SOLUTION The results from the program ADJUST are presented below.
3D Coordinate Transformation
-------------------------------------------------------J matrix
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
0.000
102.452
1284.788 1.000 0.000 0.000
⫺206.164
⫺51.103
⫺7.815
⫺195.197 0.000 1.000 0.000
⫺1355.718
⫺1287.912
195.697
4.553 0.000 0.000 1.000
53.794
0.000
118.747
1418.158 1.000 0.000 0.000
761.082
⫺62.063
28.850
723.004 0.000 1.000 0.000
⫺1496.689
⫺1421.832 ⫺722.441
42.501 0.000 0.000 1.000
65.331
0.000
129.863
1706.020 1.000 0.000 0.000
⫺1530.174
⫺61.683
⫺58.003 ⫺1451.826 0.000 1.000 0.000
⫺1799.945
⫺1709.922 1452.485
⫺41.580 0.000 0.000 1.000
64.931
0.000
204.044
1842.981 1.000 0.000 0.000
⫺50.417
⫺130.341
⫺1.911
⫺46.604 0.000 1.000 0.000
⫺1947.124
⫺1849.740
47.857
15.851 0.000 0.000 1.000
137.203
K
matrix
~~~~~~
⫺0.000
0.000
0.000
⫺0.000
⫺0.000
⫺0.000
0.060
0.209
0.000
0.033
⫺0.053
0.043
X matrix
~~~~~~~~~~~~~
⫺0.0000347107
⫺0.0000103312
⫺0.0001056763
0.1953458986
⫺0.0209088384
⫺0.0400969773
⫺0.0000257795
Measured Points
-------------------------------------------------------NAME
x
y
z
Sx
Sy
Sz
-------------------------------------------------------1
1094.883
820.085 109.821 0.007 0.008 0.005
2
503.891 1598.698 117.685 0.011 0.008 0.009
3
2349.343
207.658 151.387 0.006 0.005 0.007
4
1395.320 1348.853 215.261 0.005 0.008 0.009
TABLE 18.4 Data for a Three-Dimensional Conformal Coordinate Transformation
Point
X
Y
Z
1
2
3
4
5
6
10,037.81
10,956.68
8,780.08
10,185.80
5262.09
5128.17
4840.29
4700.21
772.04
783.00
782.62
851.32
x Sx
1094.883
503.891
2349.343
1395.320
265.346
784.081
y Sy
0.007
0.011
0.006
0.005
0.005
0.006
820.085
1598.698
207.658
1348.853
1003.470
512.683
z Sz
0.008
0.008
0.005
0.008
0.007
0.008
109.821
117.685
151.387
215.261
78.609
139.551
0.005
0.009
0.007
0.009
0.003
0.008
361
362
COORDINATE TRANSFORMATIONS
CONTROL POINTS
-------------------------------------------------------NAME
X
VX
Y
VY
Z
VZ
-------------------------------------------------------1
10037.810
0.064 5262.090
0.037 772.040
0.001
2
10956.680
0.025 5128.170 ⫺0.057 783.000
0.011
3
8780.080 ⫺0.007 4840.290 ⫺0.028 782.620
0.007
4
10185.800 ⫺0.033 4700.210
0.091 851.320 ⫺0.024
Transformation Coefficients
-------------------------Scale ⫽
0.94996
x-rot ⫽
2⬚17⬘05.3ⴖ
Y-rot ⫽ ⫺0⬚33⬘02.8ⴖ
Z-rot ⫽ 224⬚32⬘10.9ⴖ
Tx ⫽ 10233.858
Ty ⫽
6549.981
Tz ⫽
720.897
⫹/⫺
⫹/⫺
⫹/⫺
⫹/⫺
⫹/⫺
⫹/⫺
⫹/⫺
0.00004
0⬚ 00⬘ 30.1ⴖ
0⬚ 00⬘ 09.7ⴖ
0⬚ 00⬘ 06.9ⴖ
0.065
0.071
0.213
Reference Standard Deviation: 8.663
Degrees of Freedom: 5
Iterations: 2
Transformed Coordinates
-------------------------------------------------------NAME
X
Sx
Y
Sy
Z
Sz
-------------------------------------------------------1
10037.874 0.032 5262.127 0.034 772.041 0.040
2
10956.705 0.053 5128.113 0.052 783.011 0.056
3
8780.073 0.049 4840.262 0.041 782.627 0.057
4
10185.767 0.032 4700.301 0.037 851.296 0.067
5
10722.020 0.053 5691.221 0.053 766.068 0.088
6
10043.246 0.040 5675.898 0.042 816.867 0.092
Note that in this adjustment, with four control points available having X,
Y, and Z coordinates, 12 equations could be written, three for each point.
With seven unknown parameters, this gave 12 ⫺ 7 ⫽ 5 degrees of freedom
in the solution. The complete solution for this example is given in the Mathcad worksheet on the CD that accompanies this book.
18.8
STATISTICALLY VALID PARAMETERS
Besides the coordinate transformations described in preceding sections, it is
possible to develop numerous others. For example, polynomial equations of
18.8
STATISTICALLY VALID PARAMETERS
363
various degrees could be used to transform data. As additional terms are
added to a polynomial, the resulting equation will force better fits on any
given data set. However, caution should be exercised when doing this since
the resulting transformation parameters may not be statistically significant.
As an example, when using a two-dimensional conformal coordinate transformation with a data set having four control points, nonzero residuals would
be expected. However, if a projective transformation were used, this data set
would yield a unique solution, and thus the residuals would be zero. Is the
projective a more appropriate transformation for this data set? Is this truly a
better fit? Guidance in the answers to these questions can be obtained by
checking the statistical validity of the parameters.
The adjusted parameters divided by their standard deviations represent a t
statistic with degrees of freedom. If a parameter is to be judged as statistically different from zero, and thus significant, the t value computed (the test
statistic) must be greater than t␣ / 2,. Simply stated, the test statistic is
t⫽
兩parameter兩
(18.17)
S
For example, in the adjustment in Example 18.2, the following computed
t-values are found:
Parameter
a ⫽ 25.37152
b ⫽ 0.82220
c ⫽ ⫺137.183
d ⫽ ⫺0.80994
e ⫽ 25.40166
ƒ ⫽ ⫺150.723
S
t-Value
0.02532
0.02256
0.203
0.02335
0.02622
0.216
1002
36.4
675.8
34.7
968.8
697.8
In this problem there were eight equations involving six unknowns and
thus 2 degrees of freedom. From the t-distribution table (Table D.3), t0.025,2 ⫽
4.303. Because all t values computed are greater than 4.303, each parameter
is significantly different from zero at a 95% level of confidence. From the
adjustment results of Example 18.3, the t values computed are listed below.
Parameter
Value
S
t-Value
a1
b1
c1
a2
b2
c2
a3
b3
25.00274
0.80064
⫺134.715
⫺8.00771
24.99811
⫺149.815
0.00400
0.00200
0.01538
0.01896
0.377
0.00954
0.01350
0.398
0.00001
0.00002
1626
42.3
357.3
839.4
1851.7
376.4
400
100
364
COORDINATE TRANSFORMATIONS
This adjustment has eight unknown parameters and 12 observations. From
the t-distribution table (Table D.3), t0.025,4 ⫽ 2.776. By comparing the tabular t value against each computed value, all parameters are again significantly different from zero at a 95% confidence level. This is true for a3
and b3 even though they seem relatively small at 0.004 and 0.002, respectively.
Using this statistical technique, a check can be made to determine when
the projective transformation is appropriate since it defaults to an affine transformation when a3 and b3 are both statistically equal to zero. Similarly,
if the confidence intervals at a selected probability level of the twodimensional conformal coordinate transformation contain two of the parameters from the affine transformation, the computed values of the affine
transformation are statistically equal to those from the conformal transformation. Thus, if the interval for a from the conformal transformation contains both a and e from the affine transformation, there is no statistical
difference between these parameters. This must also be true for b from the
conformal transformation when compared to absolute values of b and d from
the affine transformation. Note that a negative sign is part of the conformal
coordinate transformation, and thus b and d are generally opposite in signs.
If both of these conditions exist, the conformal transformation is the more
appropriate adjustment to use for the data given. One must always be sure
that a minimum number of unknown parameters are used to solve any
problem.
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is on
within the CD supplied with the book.
18.1
Points A, B, and C have their coordinates known in both an XY and
an xy system. Points D, E, F, and G have their coordinates known
only in the xy system. These coordinates are shown in the table below.
Using a two-dimensional conformal coordinate transformation,
determine:
(a) the transformation parameters.
(b) the most probable coordinates for D, E, F, and G in the XY coordinate system.
(c) the rotation angle and scale factor.
365
PROBLEMS
18.2
Point
X
Y
x
y
A
B
C
D
E
F
G
603,462.638
604,490.074
604,314.613
390,601.450
390,987.136
391,263.879
1221.41
4607.15
4200.12
3975.00
3585.71
2767.86
2596.43
1032.09
1046.11
2946.39
1314.29
2114.28
1621.43
2692.86
Using a two-dimensional conformal coordinate transformation and the
data listed below, determine:
(a) the transformation parameters.
(b) the most probable coordinates for 9, 10, 11, and 12 in the XY
coordinate system.
(c) the rotation angle and scale factor.
Observed
Point
1
2
3
4
5
6
7
8
9
10
11
12
x
⫺4.209
14.094
⫺2.699
12.558
⫺3.930
13.805
5.743
4.146
5.584
9.809
⫺4.987
⫺0.583
Control
y
0.008
0.012
0.009
0.013
0.005
0.006
0.008
0.009
0.008
0.010
0.006
0.004
⫺6.052
9.241
10.728
⫺7.563
2.375
0.780
10.462
⫺7.288
6.493
⫺8.467
0.673
⫺5.809
0.009
0.010
0.007
0.009
0.006
0.011
0.005
0.003
0.004
0.009
0.007
0.005
X
Y
⫺106.004
⫺105.901
105.992
⫺105.967
105.697
⫺112.004
111.940
0.066
⫺0.006
106.155
105.939
⫺105.991
⫺0.024
⫺0.108
112.087
⫺111.991
18.3
Do Problem 18.2 using an unweighted least squares adjustment.
18.4
Do parts (a) and (b) in Problem 18.2 using a two-dimensional affine
coordinate transformation.
18.5
Do parts (a) and (b) in Problem 18.2 using a two-dimensional projective coordinate transformation.
18.6
Determine the appropriate two-dimensional transformation for Problem 18.2 at a 0.01 level of significance.
366
COORDINATE TRANSFORMATIONS
18.7
Using a two-dimensional affine coordinate transformation and the following data, determine:
(a) the transformation parameters.
(b) the most probable XY coordinates for points 9 to 12.
Point
x
1
2
3
4
5
6
7
8
9
10
11
12
⫺83.485 0.005
⫺101.331 0.006
⫺43.818 0.011
16.737 0.015
42.412 0.006
60.360 0.010
2.788 0.006
⫺57.735 0.003
⫺63.048 0.008
45.103 0.007
⫺7.809 0.004
57.309 0.008
y
1.221
56.123
38.462
13.140
⫺44.813
⫺99.889
⫺82.065
⫺56.556
⫺89.056
32.887
98.773
⫺17.509
0.007
0.010
0.012
0.013
0.009
0.008
0.012
0.005
0.008
0.006
0.010
0.009
X
Y
⫺113.000
⫺105.962
0.003
105.598
112.993
105.996
⫺0.002
⫺105.934
⫺112.986
⫺105.628
0.001
105.998
112.884
105.889
⫺0.001
⫺105.887
18.8
Do Problem 18.7 using an unweighted least squares adjustment.
18.9
Do Problem 18.7 using a two-dimensional projective coordinate transformation.
18.10 For the data of Problem 18.7, which two-dimensional transformation
is most appropriate, and why? Use a 0.01 level of significance.
18.11 Determine the appropriate two-dimensional coordinate transformation
for the following data at a 0.01 level of significance.
Point
X (m)
Y (m)
x (mm)
y (mm)
x
y
1
2
3
4
5
6
2181.578
1145.486
⫺855.426
1087.225
2540.778
⫺2595.242
2053.274
809.022
383.977
⫺1193.347
⫺2245.477
1926.548
89.748
49.942
⫺29.467
50.164
109.599
⫺100.971
91.009
39.960
20.415
⫺40.127
⫺80.310
79.824
0.019
0.016
0.028
0.028
0.018
0.026
0.020
0.021
0.028
0.028
0.021
0.022
18.12 Using a weighted three-dimensional conformal coordinate transformation, determine the transformation parameters for the following
data set.
367
PROBLEMS
Point
X (m)
Y (m)
1
2
3
4
8948.16
8813.93
8512.60
8351.02
6678.50
5755.23
7937.11
6483.62
Z (m)
x (mm)
y (mm) z (mm)
x
756.51 1094.97 810.09 804.73 0.080
831.67 508.31 1595.68 901.78 0.080
803.11 2356.23 197.07 834.47 0.097
863.24 1395.18 1397.64 925.96 0.043
y
z
0.084
0.060
0.177
0.161
0.153
0.069
0.202
0.120
18.13 Do Problem 18.12 using an unweighted least squares adjustment.
18.14 Using a weighted three-dimensional conformal coordinate transformation and the follow set of data:
(a) determine the transformation parameters.
(b) Compute the XYZ coordinates for points 7 to 10.
Control points
Point
X (m)
Y (m)
Z (m)
1
2
3
4
5
6
9770.192
16371.750
5417.336
27668.765
—
—
16944.028
14998.190
265.432
26963.937
—
—
1235.280
1407.694
—
—
1325.885
1070.226
Measured points
Point
x (m)
1
9845.049 0.015
2
16,441.006 0.015
3
4
5433.174 0.015
27781.044 0.015
5
8543.224 0.015
6
4140.096 0.015
7
8
23,125.031 0.015
4893.721 0.015
9
10
19967.763 0.015
2569.022 0.015
y (m)
z (m)
16,911.947
0.015
14,941.872
0.015
250.766 0.0115
26864.597
0.015
22,014.402
0.015
24,618.211
0.015
4672.275 0.015
12,668.887
0.015
1603.499 0.015
14,610.600
0.015
1057.242 0.025
1169.148 0.025
1476.572 0.025
861.956 0.025
1139.204 0.025
918.253 0.025
1351.655 0.025
1679.184 0.025
1210.986 0.025
1359.663 0.025
368
COORDINATE TRANSFORMATIONS
18.15 Do Problem 11.19, and determine whether the derived constant and
scale factor are statistically significant at a 0.01 level of significance.
Use the program ADJUST to do each problem.
18.16 Problem 18.6
18.17 Problem 18.10
18.18 Problem 18.14
Programming Problems
Develop a computational program that calculates the coefficient and constants
matrix for each transformation.
18.19 A two-dimensional conformal coordinate transformation
18.20 A two-dimensional affine coordinate transformation
18.21 A two-dimensional projective coordinate transformation
18.22 A three-dimensional conformal coordinate transformation
CHAPTER 19
ERROR ELLIPSE
19.1
INTRODUCTION
As discussed previously, after completing a least squares adjustment, the estimated standard deviations in the coordinates of an adjusted station can be
calculated from covariance matrix elements. These standard deviations provide error estimates in the reference axes directions. In graphical representation, they are half the dimensions of a standard error rectangle centered on
each station. The standard error rectangle has dimensions of 2Sx by 2Sy as
illustrated for station B in Figure 19.1, but this is not a complete representation
of the error at the station.
Simple deductive reasoning can be used to show the basic problem. Assume in Figure 19.1 that the XY coordinates of station A have been computed
from the observations of distance AB and azimuth AzAB that is approximately
30⬚. Further assume that the observed azimuth has no error at all but that the
distance has a large error, say 2 ft. From Figure 19.1 it should then be
readily apparent that the largest uncertainty in the station’s position would
not lie in either cardinal direction. That is, neither Sx nor Sy represents the
largest positional uncertainty for the station. Rather, the largest uncertainty
would be collinear with line AB and approximately equal to the estimated
error in the distance. In fact, this is what happens.
In the usual case, the position of a station is uncertain in both direction
and distance, and the estimated error of the adjusted station involves the errors
of two jointly distributed variables, the x and y coordinates. Thus, the positional error at a station follows a bivariate normal distribution. The general
shape of this distribution for a station is shown in Figure 19.2. In this figure,
the three-dimensional surface plot [Figure 19.2(a)] of a bivariate normal disAdjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
369
370
ERROR ELLIPSE
Figure 19.1 Standard error rectangle.
tribution curve is shown along with its contour plot [Figure 19.2(b)]. Note
that the ellipses shown in the xy plane of Figure 19.2(b) can be obtained by
passing planes of varying levels through Figure 19.2(a) parallel to the xy
plane. The volume of the region inside the intersection of any plane with the
surface of Figure 19.2(a) represents the probability level of the ellipse. The
orthogonal projection of the surface plot of Figure 19.2(a) onto the xz plane
would give the normal distribution curve of the x coordinate, from which Sx
is obtained. Similarly, its orthogonal projection onto the yz plane would give
the normal distribution in the y coordinate from which Sy is obtained.
To fully describe the estimated error of a station, it is only necessary to
show the orientation and lengths of the semiaxes of the error ellipse. A detailed diagram of an error ellipse is shown in Figure 19.3. In this figure, the
standard error ellipse of a station is shown (i.e., one whose arcs are tangent
Figure 19.2 (a) Three-dimensional view and (b) contour plot of a bivariate normal
distribution.
19.2
COMPUTATION OF ELLIPSE ORIENTATION AND SEMIAXES
371
Figure 19.3 Standard error ellipse.
to the sides of the standard error rectangle). The orientation of the ellipse
depends on the t angle, which fixes the directions of the auxiliary, orthogonal
uv axes along which the ellipse axes lie. The u axis defines the weakest
direction in which the station’s adjusted position is known. In other words, it
lies in the direction of maximum error in the station’s coordinates. The v axis
is orthogonal to u and defines the strongest direction in which the station’s
position is known, or the direction of minimum error. For any station, the
value of t that orients the ellipse to provide these maximum and minimum
values can be determined after the adjustment from the elements of the covariance matrix.
The exact probability level of the standard error ellipse is dependent on
the number of degrees of freedom in the adjustment. This standard error
ellipse can be modified in dimensions through the use of F statistical values
to represent an error probability at any percentage selected. It will be shown
later that the percent probability within the boundary of standard error ellipse
for a simple closed traverse is only 35%. Surveyors often use the 95% probability since this affords a high level of confidence.
19.2
COMPUTATION OF ELLIPSE ORIENTATION AND SEMIAXES
As shown in Figure 19.4, the method for calculating the orientation angle t
that yields maximum and minimum semiaxes involves a two-dimensional co-
Figure 19.4 Two-dimensional rotation.
372
ERROR ELLIPSE
ordinate rotation. Notice that the t angle is defined as a clockwise angle from
the y axis to the u axis. To propagate the errors in a point I from the xy
system into an orthogonal uv system, the generalized law of the propagation
of variances, discussed in Chapter 6, is used. The specific value for t that
yields the maximum error along the u axis must be determined. The following
steps accomplish this task.
Step 1: Any point I in the uv system can be represented with respect to its
xy coordinates as
ui ⫽ xi sin t ⫹ yi cos t
(19.1)
vi ⫽ ⫺xi cos t ⫹ yi sin t
Rewriting Equations (19.1) in matrix form yields
冋册 冋
ui
vi
⫽
册冋 册
sin t cos t
sin t
⫺cos t
xi
yi
(19.2)
or in simplified matrix notation,
(19.3)
Z ⫽ RX
Step 2: Assume that for the adjustment problem in which I appears, there is
a Qxx matrix for the xy coordinate system. The problem is to develop, from
the Qxx matrix, a new covariance matrix Qzz for the uv coordinate system.
This can be done using the generalized law for the propagation of variances, given in Chapter 6 as
⌺zz ⫽ S 02 RQxx RT
(19.4)
where
R⫽
冋
册
sin t cos t
sin t
⫺cos t
Since S 20 is a scalar, it can be dropped temporarily and recalled again
after the derivation. Thus,
Qzz ⫽ RQxx RT ⫽
where
冋
quu
quv
册
quv
qvv
(19.5)
19.2
373
COMPUTATION OF ELLIPSE ORIENTATION AND SEMIAXES
Qxx ⫽
冋
册
qxx qxy
qxy qyy
Step 3: Expanding Equation (19.5), the elements of the Qzz matrix are
Qzz ⫽
冉
冉
冤
冊 冉
冊 冉
冊
冊
qxx sin2 t ⫹ qxy cos t sin t
⫹ qxy sin t cos t ⫹ qyy cos2 t
⫺qxx sin t cos t ⫺ qxy cos2 t
⫹qxy sin2 t ⫹ qyy cos t sin t
⫺qxx cos t sin t ⫹ qxy sin2 t
⫺qxy cos2 t ⫹ qyy sin t cos t
qxx cos2 t ⫺ qxy sin t cos t
⫺qxy cos t sin t ⫹ qyy sin2 t
⫽
冋
册
quu quv
quv qvv
冥
(19.6)
Step 4: The quu element of Equation (19.6) can be rewritten as
quu ⫽ qxx sin2 t ⫹ 2qxy cos t sin t ⫹ qyy cos2 t
(19.7)
The following trigonometric identities are useful in developing an equation
for t:
sin 2t ⫽ 2 sin t cos t
(a)
cos 2t ⫽ cos2 t ⫺ sin2 t
(b)
cos2 t ⫹ sin2 t ⫽ 1
(c)
Substituting identity (a) into Equation (19.7) yields
quu ⫽ qxx sin2 t ⫹ qyy cos2 t ⫹ 2qxy
sin 2t
2
(19.8)
Regrouping the first two terms and adding the necessary terms to Equation
(19.8) to maintain equality yields
quu ⫽
qxx ⫹ qyy
q sin2 t qyy cos2 t
(sin2 t ⫹ cos2 t) ⫹ xx
⫹
2
2
2
⫺
qyy sin2 t qxx cos2 t
⫺
⫹ qxy sin 2t
2
2
Substituting identity (c) and regrouping Equation (19.9) results in
(19.9)
374
ERROR ELLIPSE
quu ⫽
qxx ⫹ qyy qyy
⫹
(cos2 t ⫺ sin2 t)
2
2
⫺
qxx
(cos2 t ⫺ sin2 t) ⫹ qxy sin 2t
2
(19.10)
Finally, substituting identity (b) into Equation (19.10) yields
quu ⫽
qxx ⫹ qyy qyy ⫺ qxx
⫹
cos 2t ⫹ qxy sin 2t
2
2
(19.11)
To find the value of t that maximizes quu, differentiate quu in Equation
(19.8) with respect to t and set the results equal to zero. This results in
qyy ⫺ qxx
dquu
⫽⫺
2 sin 2t ⫹ 2qxy cos 2t ⫽ 0
dt
2
(19.12)
Reducing Equation (19.12) yields
(qyy ⫺ qxx) sin 2t ⫽ 2qxy cos 2t
(19.13)
Finally, dividing Equation (19.13) by cos 2t yields
2qxy
sin 2t
⫽ tan 2t ⫽
cos 2t
qyy ⫺ qxx
(19.14a)
Equation (19.14a) is used to compute 2t and hence the desired angle t
that yields the maximum value of quu. Note that the correct quadrant of 2t
must be determined by noting the sign of the numerator and denominator
in Equation (19.14a) before dividing by 2 to obtain t. Table 19.1 shows
the proper quadrant for the different possible sign combinations of the
TABLE 19.1 Selection of the Proper Quadrant
for 2t a
Algebraic Sign of
sin 2t
cos 2t
Quadrant
⫹
⫹
⫺
⫺
⫹
⫺
⫺
⫹
1
2
3
4
a
When calculating for t, always remember to select the proper
quadrant of 2t before dividing by 2.
19.2
COMPUTATION OF ELLIPSE ORIENTATION AND SEMIAXES
375
numerator and denominator.
Table 19.1 can be avoided by using the atan2 function available in most
software packages. This function returns a value between ⫺180⬚ ⱕ 2t ⱕ
180⬚. If the value returned is negative, the correct value for 2t is obtained
by adding 360⬚. Correct use of the atan2 function is
2t ⫽ atan 2(qyy ⫺ qxx, 2qxy)
(19.14b)
The use of Equation (19.14b) is demonstrated in the Mathcad worksheet
for this chapter on the CD that accompanies this book.
Correlation between the latitude and departure of a station was discussed
in Chapter 8. Similarly, the adjusted coordinates of a station are also correlated. Computing the value of t that yields the maximum and minimum
values for the semiaxes is equivalent to rotating the covariance matrix until
the off-diagonals are nonzero. Thus, the u and v coordinate values will be
uncorrelated, which is equivalent to setting the quv element of Equation
(19.6) equal to zero. Using the trigonometric identities noted previously,
the element quv from Equation (19.6) can be written as
quv ⫽
qxx ⫺ qyy
sin 2t ⫹ qxy cos 2t
2
(19.15)
Setting quv equal to zero and solving for t gives us
2qxy
sin 2t
⫽ tan 2t ⫽
cos 2t
qyy ⫺ qxx
(19.16)
Note that this yields the same result as Equation (19.14), which verifies
the removal of the correlation.
Step 5: In a fashion similar to that demonstrated in step 4, the qvv element
of Equation (19.6) can be rewritten as
qvv ⫽ qxx cos2 t ⫺ 2qxy cos t sin t ⫹ qyy sin2 t
(19.17)
In summary, the t angle, semimajor ellipse axis (quu), and semiminor axis
(qvv) are calculated using Equations (19.14), (19.7), and (19.17), respectively.
These equations are repeated here, in order, for convenience. Note that these
equations use only elements from the covariance matrix.
tan 2t ⫽
2qxy
qyy ⫺ qxx
quu ⫽ qxx sin2 t ⫹ 2qxy cos t sin t ⫹ qyy cos2 t
(19.18)
(19.19)
376
ERROR ELLIPSE
qvv ⫽ qxx cos2 t ⫺ 2qxy cos t sin t ⫹ qyy sin2 t
(19.20)
Equation (19.18) gives the t angle that the u axis makes with the y axis.
Equation (19.19) yields the numerical value for quu, which when multiplied
by the reference variance S 20 gives the variance along the u axis. The square
root of the variance is the semimajor axis of the standard error ellipse. Equation (19.20) yields the numerical value for qvv, which when multiplied by
S 20 gives the variance along the v axis. The square root of this variance is the
semiminor axis of the standard error ellipse. Thus, the semimajor and semiminor axes are
Su ⫽ S0quu and Sv ⫽ S0qvv
(19.21)
19.3 EXAMPLE PROBLEM OF STANDARD ERROR
ELLIPSE CALCULATIONS
In this section the error ellipse data for the trilateration example in Section
14.5 are calculated. From the computer listing given for the solution of that
problem, the following values are recalled:
1. S0 ⫽ 0.136 ft
2. The unknown X and covariance Qxx matrices were
冤 冥
冤
dXWisconsin
dYWisconsin
X⫽
dXCampus
dYCampus
Qxx ⫽
19.3.1
冥
1.198574 ⫺1.160249 ⫺0.099772 ⫺1.402250
2.634937
0.193956
2.725964
0.193956
0.583150
0.460480
2.725964
0.460480
3.962823
⫺1.160249
⫺0.099772
⫺1.402250
Error Ellipse for Station Wisconsin
The tangent of 2t is
tan 2t ⫽
2(⫺1.160249)
⫽ ⫺1.6155
2.634937 ⫺ 1.198574
Note that the sign of the numerator is negative and the denominator is
positive. Thus, from Table 19.1, angle 2t is in the fourth quadrant and 360⬚
must be added to the computed angle. Hence,
19.3
EXAMPLE PROBLEM OF STANDARD ERROR ELLIPSE CALCULATIONS
377
2t ⫽ tan⫺1(⫺1.6155) ⫽ ⫺58⬚14.5⬘ ⫹ 360⬚ ⫽ 301⬚45.5⬘
t ⫽ 150⬚53⬘
Substituting the appropriate values into Equation (19.21), Su is
Su ⫽ 0.136
⫻ 1.198574 sin2 t ⫹ 2(⫺1.160249) cos t sin t ⫹ 2.634937 cos2 t
⫽ 0.25 ft
Similarly substituting the appropriate values into Equation (19.21), Sv is
Sv ⫽ 0.136
⫻ 1.198574 cos2 t ⫺ 2(⫺1.160249) cos t sin t ⫹ 2.634937 sin2 t
⫽ 0.10 ft
Note that the standard deviations in the coordinates, as computed by Equation
(13.24), are
Sx ⫽ S0qxx ⫽ 0.1361.198574 ⫽ 0.15 ft
Sy ⫽ S0qyy ⫽ 0.1362.634937 ⫽ 0.22 ft
19.3.2
Error Ellipse for Station Campus
Using the same procedures as in Section 19.3.1, the error ellipse data for
station Campus are
2t ⫽ tan⫺1
2 ⫻ 0.460480
⫽ 15⬚14⬘
3.962823 ⫺ 0.583150
t ⫽ 7⬚37⬘
Su ⫽ 0.136
⫻ 0.583150 sin2 t ⫹ 2(0.460480) cos t sin t ⫹ 3.962823 cos2 t
⫽ 0.27 ft
Sv ⫽ 0.136
⫻ 0.583150 cos2 t ⫺ 2(0.460480) cos t sin t ⫹ 3.962823 sin2 t
⫽ 0.10 ft
378
ERROR ELLIPSE
Sx ⫽ S0qxx ⫽ 0.1360.583150 ⫽ 0.10 ft
Sy ⫽ S0qyy ⫽ 0.1363.962823 ⫽ 0.27 ft
19.3.3
Drawing the Standard Error Ellipse
To draw the error ellipses for stations Wisconsin and Campus of Figure 19.5,
the error rectangle is first constructed by laying out the values of Sx and Sy
using a convenient scale along the x and y adjustment axes, respectively. For
this example, an ellipse scale of 4800 times the map scale was selected. The
t angle is laid off clockwise from the positive y axis to construct the u axis.
The v axis is drawn 90⬚ counterclockwise from u to form a right-handed
coordinate system. The values of Su and Sv are laid off along the U and V
axes, respectively, to locate the semiaxis points. Finally, the ellipse is constructed so that it is tangent to the error rectangle and passes through its
semiaxes points (refer to Figure 19.3).
19.4
ANOTHER EXAMPLE PROBLEM
In this section, the standard error ellipse for station u in the example of
Section 16.4 is calculated. For the adjustment, S0 ⫽ 1.82 ft, and the X and
Qxx matrices are
X⫽
冋 册
dxu
dyu
Qxx ⫽
冋
册 冋
册
qxx qxy
0.000532 0.000602
⫽
qxy qyy
0.000602 0.000838
Error ellipse calculations are
Figure 19.5 Graphical representation of error ellipses.
19.5
2t ⫽ tan⫺1
ERROR ELLIPSE CONFIDENCE LEVEL
379
2 ⫻ 0.000602
⫽ 75⬚44⬘
0.000838 ⫺ 0.000532
t ⫽ 37⬚52⬘
Su ⫽ 1
⫻ 0.000532 sin2 t ⫹ 2(0.000602) cos t sin t ⫹ 0.000838 cos2 t
⫽ 0.036 ft
Sv ⫽ 1
⫻ 0.000532 cos2 t ⫺ 2(0.000602) cos t sin t ⫹ 0.000838 sin2 t
⫽ 0.008 ft
Sx ⫽ 10.000532 ⫽ 0.023 ft
Sy ⫽ 10.000838 ⫽ 0.029 ft
Figure 19.6 shows the plotted standard error ellipse and its error rectangle.
19.5
ERROR ELLIPSE CONFIDENCE LEVEL
The calculations in Sections 19.3 and 19.4 produce standard error ellipses.
These ellipses can be modified to produce error ellipses at any confidence
level ␣ by using an F statistic with two numerator degrees of freedom and
the degrees of freedom for the adjustment in the denominator. Since the F
statistic represents variance ratios for varying degrees of freedom, it can be
Figure 19.6 Graphical representation of error ellipse.
380
ERROR ELLIPSE
TABLE 19.2 F␣,2,degrees of freedom Statistics for Selected Probability Levels
Probability
Degrees of Freedom
90%
95%
99%
1
2
3
4
5
10
15
20
30
60
49.50
9.00
5.46
4.32
3.78
2.92
2.70
2.59
2.49
2.39
199.5
19.00
9.55
6.94
5.79
4.10
3.68
3.49
3.32
3.15
4999.50
99.00
30.82
18.00
13.27
7.56
6.36
5.85
5.39
4.98
expected that with increases in the number of degrees of freedom, there will
be corresponding increases in precision. The F(␣,2,degrees of freedom) statistic modifier for various confidence levels is listed in Table 19.2. Notice that as the
degrees of freedom increase, the F statistic modifiers decrease rapidly and
begin to stabilize for larger degrees of freedom. The confidence level of an
error ellipse can be increased to any level by using the multiplier
c ⫽ 2F(␣,2,degrees
of freedom)
(19.22)
Using the following equations, the percent probability for the semimajor and
semiminor axes can be computed as
Su% ⫽ Suc ⫽ Su2F(␣,2,degrees
of freedom)
Sv% ⫽ Svc ⫽ Sv2F(␣,2,degrees
of freedom)
From the foregoing it should be apparent that as the number of degrees of
freedom (redundancies) increases, precision increases, and the error ellipse
sizes decrease. Using the techniques discussed in Chapter 4, the values listed
in Table 19.2 are for the F distribution at 90% (␣ ⫽ 0.10), 95% (␣ ⫽ 0.05),
and 99% (␣ ⫽ 0.01) probability. These probabilities are most commonly used.
Values from this table can be substituted into Equation (19.22) to determine
the value of c for the probability selected and the given number of redundancies in the adjustment. This table is for convenience only and does not contain
the values necessary for all situations that might arise.
Example 19.1 Calculate the 95% error ellipse for station Wisconsin of Section 19.3.1.
19.6
ERROR ELLIPSE ADVANTAGES
381
SOLUTION Using Equations (19.23) yields
Su95% ⫽ 0.252(199.50) ⫽ 4.99 ft
Sv95% ⫽ 0.102(199.50) ⫽ 2.00 ft
Sx95% ⫽ 2(199.50)Sx ⫽ (19.97 ⫻ 0.15) 3.00 ft
Sy95% ⫽ 2(199.50)Sy ⫽ (19.97 ⫻ 0.22) 4.39 ft
The probability of the standard error ellipse can be found by setting the
multiplier 2F␣,1,2 equal to 1 so that F␣,1,2 ⫽ 0.5. For a simple closed traverse
with three degrees of freedom, this means that F␣,2,3 ⫽ 0.5. The value of ␣
that satisfies this condition is 0.65 and was found by trial-and-error procedures
using the program STATS. Thus, the percent probability of the standard error
ellipse in a simple closed traverse is (1 ⫺ 0.65) ⫻ 100%, or 35%. The reader
is encouraged to verify this result using program STATS. It is left as an
exercise for the reader to show that the percent probability for the standard
error ellipse ranges from 35% to only 39% for horizontal surveys that have
fewer than 100 degrees of freedom.
19.6
ERROR ELLIPSE ADVANTAGES
Besides providing critical information regarding the precision of an adjusted
station position, a major advantage of error ellipses is that they offer a method
of making a visual comparison of the relative precisions between any two
stations. By viewing the shapes, sizes, and orientations of error ellipses, various surveys can be compared rapidly and meaningfully.
19.6.1
Survey Network Design
The sizes, shapes, and orientations of error ellipses depend on (1) the control
used to constrain the adjustment, (2) the observational precisions, and (3) the
geometry of the survey. The last two of these three elements are variables
that can be altered in the design of a survey to produce optimal results. In
designing surveys that involve the traditional observations of distances and
angles, estimated precisions can be computed for observations made with
differing combinations of equipment and field procedures. Also, trial variations in station placement, which creates the network geometry, can be made.
Then these varying combinations of observations and geometry can be pro-
382
ERROR ELLIPSE
cessed through least squares adjustments and the resulting station error ellipses computed, plotted, and checked against the desired results. Once
acceptable goals are achieved in this process, the observational equipment,
field procedures, and network geometry that provide these results can be
adopted. This overall process is called network design. It enables surveyors
to select the equipment and field techniques, and to decide on the number
and locations of stations that provide the highest precision at lowest cost. This
can be done in the office using simulation software before bidding a contract
or entering the field.
In designing networks to be surveyed using the traditional observations of
distance, angle, and directions, it is important to understand the relationships
of those observations to the resulting positional uncertainties of the stations.
The following relationships apply:
1. Distance observations strengthen the positions of stations in directions
collinear with the lines observed.
2. Angle and direction observations strengthen the positions of stations in
directions perpendicular to the lines of sight.
A simple analysis made with reference to Figure 19.1 should clarify the
two relationships above. Assume first that the length of line AB was measured
precisely but its direction was not observed. Then the positional uncertainty
of station B should be held within close tolerances by the distance observed,
but it would only be held in the direction collinear with AB. The distance
observation would do nothing to keep line AB from rotating, and in fact the
position of B would be weak perpendicular to AB. On the other hand, if the
direction of AB had been observed precisely but its length had not been
measured, the positional strength of station B would be strongest in the direction perpendicular to AB. But an angle observation alone does nothing to
fix distances between observed stations, and thus the position of station B
would be weak along line AB. If both the length and direction AB were
observed with equal precision, a positional uncertainty for station B that is
more uniform in all directions would be expected. In a survey network that
consists of many interconnected stations, analyzing the effects of observations
is not quite as simple as was just demonstrated for the single line AB. Nevertheless, the two basic relationships stated above still apply.
Uniform positional strength in all directions for all stations is the desired
goal in survey network design. This would be achieved if, following least
squares adjustment, all error ellipses were circular in shape and of equal size.
Although this goal is seldom completely possible, by diligently analyzing
various combinations of geometric figures together with different combinations of observations and precisions, this goal can be approached. Sometimes,
however, other overriding factors, such as station accessibility, terrain, and
vegetation, preclude actual use of an optimal design.
19.6
ERROR ELLIPSE ADVANTAGES
383
The network design process discussed above is significantly aided by the
use of aerial photos and/or topographic maps. These products enable the
layout of trial station locations and permits analysis of the accessibility and
intervisibility of these stations to be investigated. However, a field reconnaissance should be made before adopting the final design.
The global positioning system (GPS) has brought about dramatic changes
in all areas of surveying, and network design is not an exception. Although
GPS does require overhead visibility at each receiver station for tracking
satellites, problems of intervisibility between ground stations is eliminated.
Thus, networks having uniform geometry can normally be laid out. Because
each station in the network is occupied in a GPS survey, the XYZ coordinates
of the stations can be determined. This simplifies the problem of designing
networks to attain error ellipses of uniform shapes and sizes. However, the
geometric configuration of satellites is an important factor that affects station
precisions. The positional dilution of precision (PDOP) can be a guide to the
geometric strength of the observed satellites. In this case, the lower the PDOP,
the stronger the satellite geometry. For more discussion on designing GPS
surveys, readers are referred to books devoted to the subject of GPS surveying.
19.6.2
Example Network
Figure 19.7 shows error ellipses for two survey networks. Figure 19.7(a)
illustrates the error ellipses from a trilateration survey with the nine stations,
Figure 19.7 Network analysis using error ellipses: (a) trilateration for 19 distances;
(b) triangulation for 19 angles.
384
ERROR ELLIPSE
two of which (Red and Bug) were control stations. The survey includes 19
distance observations and five degrees of freedom. Figure 19.7(b) shows the
error ellipses of the same network that was observed using triangulation and
a baseline from stations Red to Bug. This survey includes 19 observed angles
and thus also has five degrees of freedom.
With respect to these two figures, and keeping in mind that the smaller the
ellipse, the higher the precision, the following general observations can be
made:
1. In both figures, stations Sand and Birch have the highest precisions.
This, of course, is expected due to their proximity to control station
Bug and because of the density of observations made to these stations,
which included direct measurements from both control stations.
2. The large size of error ellipses at stations Beaver, Schutt, Bunker, and
Bee of Figure 19.7(b) show that they have lower precision. This, too,
is expected because there were fewer observations made to those stations. Also, neither Beaver nor Bee was connected directly by observation to either of the control stations.
3. Stations White and Schutt of Figure 19.7(a) have relatively high east–
west precisions and relatively low north–south precisions. Examination
of the network geometry reveals that this could be expected. Distance
measurements to those two points from station Red, plus an observed
distance between White and Schutt, would have greatly improved the
north–south precision.
4. Stations Beaver and Bunker of Figure 19.7(a) have relatively low precisions east–west and relatively high precisions north–south. Again, this
is expected when examining the network geometry.
5. The smaller error ellipses of Figure 19.7(a) suggest that the trilateration
survey will yield superior precision to the triangulation survey of Figure
19.7(b). This is expected since the EDM had a stated uncertainty of
(5 mm ⫹ 5 ppm). In a 5000-ft distance this yields an uncertainty of
0.030 ft. To achieve the same precision, the comparable angle uncertainty would need to be
ⴖ ⫽
S
0.030
⫽
206,264.8ⴖ /rad ⫽ 1.2ⴖ
R
5000
The proposed instrument and field procedures for the project that
yielded the error ellipses of Figure 19.7(b) had an expected uncertainty
of only 6ⴖ. Very probably, this ultimate design would include both
observed distances and angles.
These examples serve to illustrate the value of computing station error
ellipses in an a priori analysis. The observations were made easily and quickly
19.7
OTHER MEASURES OF STATION UNCERTAINTY
385
TABLE 19.3 Other Measures of Two-Dimensional
Positional Uncertainties
Probability (%)
c
Name
35–39
50.0
63.2
86.5
95.0
98.2
98.9
1.00
1.18
1.41
2.00
2.45
2.83
3.00
Standard error ellipse
Circular error probable (CEP)
Distance RMS (DRMS)
Two-sigma ellipse
95% confidence level
2DRMS
Three-sigma ellipse
by comparison of the ellipses in the two figures. Similar information would
have been difficult, if not impossible, to determine from standard deviations.
By varying the survey it is possible ultimately to find a design that provides
optimal results in terms of meeting a uniformly acceptable precision and
survey economy.
19.7
OTHER MEASURES OF STATION UNCERTAINTY
Other measures of accuracies are sometimes called for in specifications. As
discussed in Section 19.5, the standard error ellipse has a c-multiplier of 1.00
and a probability between 35 and 39%. Other common errors and probabilities
are given in Table 19.3.
As demonstrated in the Mathcad worksheet on the CD that accompanies
this book, the process of rotating the 2 ⫻ 2 block diagonal matrix for a station
is the mathematical equivalent of orthogonalization. This process can be performed by computing eigenvalues and eigenvectors. For example, the eigenvalues of the 2 ⫻ 2 block diagonal matrix for station Wisconsin in Example
19.2 are 0.55222 and 3.28129, respectively. Thus, SU-Wis is 0.1363.28129
⫽ 0.25 ft and SV-Wis is 0.1360.55222 ⫽ 0.101 ft.
TABLE 19.4 Measures of Three-Dimensional Positional Uncertainties
Probability (%)
c
Name
19.9
50.0
61.0
73.8
95.0
97.1
1.00
1.53
1.73
2.00
2.80
3.00
Standard ellipsoid
Spherical error probable (SEP)
Mean radical spherical error (MRSE)
Two-sigma ellipsoid
95% confidence level
Three-sigma ellipsoid
386
ERROR ELLIPSE
This property can be used to compute the error ellipsoids for threedimensional coordinates from a GPS adjustment or the three-dimensional geodetic network adjustment discussed in Chapter 23. That is, the uncertainties
along the three orthogonal axes of the error ellipsoid can be computed using
eigenvalues of the 3 ⫻ 3 block diagonal matrix appropriate for each station.
The common measures for ellipsoids are listed in Table 19.4.
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
19.1
Calculate the semiminor and semimajor axes of the standard error
ellipse for the adjusted position of station U in Example 14.2. Plot
the figure using a scale of 1⬊12,000 and the error ellipse using an
appropriate scale.
19.2
Calculate the semiminor and semimajor axes of the 95% confidence
error ellipse for Problem 19.1. Plot this ellipse superimposed over the
ellipse of Problem 19.1.
19.3
Repeat Problem 19.1 for Example 15.1.
19.4
Repeat Problem 19.2 for Problem 19.3.
19.5
Repeat Problem 19.1 for the adjusted position of stations B and C of
Example 15.3. Use a scale of 1⬊24,000 for the figure and plot the
error ellipses using an appropriate multiplication factor.
Calculate the error ellipse data for the unknown stations in each problem.
19.6
Problem 15.2
19.7
Problem 15.5
19.8
Problem 15.6
19.9
Problem 15.9
19.10 Problem 15.12
19.11 Problem 16.2
19.12 Problem 16.3
19.13 Problem 16.7
19.14 Problem 16.9
PROBLEMS
387
Using a level of significance of 0.05, compute the 95% probable error ellipse
for the stations in each problem.
19.15 Problem 15.2
19.16 Problem 16.3
19.17 Using the program STATS, determine the percent probability of the
standard error ellipse for a horizontal survey with:
(a) 3 degrees of freedom.
(b) 9 degrees of freedom.
(c) 20 degrees of freedom.
(d) 50 degrees of freedom.
(e) 100 degrees of freedom.
Programming Problems
19.18 Develop a computational program that takes the Qxx matrix and S 20
from a horizontal adjustment and computes error ellipse data for the
unknown stations.
19.19 Develop a computational program that does the same as described for
Problems 19.15 and 19.16.
CHAPTER 20
CONSTRAINT EQUATIONS
20.1
INTRODUCTION
When doing an adjustment, it is sometimes necessary to fix an observation
to a specific value. For instance, in Chapter 14 it was shown that the coordinates of a control station can be fixed by setting its dx and dy corrections
to zero, and thus the corrections and their corresponding coefficients in the J
matrix were removed from the solution. This is called a constrained adjustment. Another constrained adjustment occurs when the direction or length of
a line is held to a specific value or when an elevation difference between two
stations is fixed in differential leveling. In this chapter methods available for
developing observational constraints are discussed. However, before discussing constraints, the procedure for including control station coordinates in an
adjustment is described.
20.2
ADJUSTMENT OF CONTROL STATION COORDINATES
In examples in preceding chapters, when control station coordinates were
excluded from the adjustments and hence their values were held fixed, constrained adjustments were being performed. That is, the observations were
being forced to fit the control coordinates. However, control is not perfect
and not all control is of equal reliability. This is evidenced by the fact that
different orders of accuracy are used to classify control.
When more than minimal control is held fixed in an adjustment, the observations are forced to fit this control. For example, if the coordinates of two
control stations are held fixed but their actual positions are not in agreement
with the values given by their held coordinates, the observations will be adjusted to match the erroneous control. Simply stated, precise observations may
be forced to fit less precise control. This was not a major problem in the days
388
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
20.2
ADJUSTMENT OF CONTROL STATION COORDINATES
389
of transits and tapes, but it does happen with modern instrumentation. This
topic is discussed in more detail in Chapter 21.
To clarify the problem further, suppose that a new survey is tied to two
existing control stations set from two previous surveys. Assume that the precision of the existing control stations is only 1⬊10,000. A new survey uses
equipment and field procedures designed to produce a survey of 1⬊20,000.
Thus, it will have a higher accuracy than either of the control stations to
which it must fit. If both existing control stations are fixed in the adjustment,
the new observations must distort to fit the errors of the existing control
stations. After the adjustment, their residuals will show a lower-order fit that
matches the control. In this case it would be better to allow the control coordinates to adjust according to their assigned quality so that the observations
are not distorted. However, it should be stated that the precision of the new
coordinates relative to stations not in the adjustment can be only as good as
the initial control.
The observation equations for control station coordinates are
x⬘ ⫽ x ⫹ vx
(20.1)
y⬘ ⫽ y ⫹ vy
In Equation (20.1) x⬘ and y⬘ are the observed coordinate values of the control
station, x and y the published coordinate values of the control station, and vx
and vy the residuals for the respective published coordinate values.
To allow the control to adjust, Equations (20.1) must be included in the
adjustment for each control station. To fix a control station in this scheme,
high weights are assigned to the station’s coordinates. Conversely, low
weights will allow a control station’s coordinates to adjust. In this manner,
all control stations are allowed to adjust in accordance with their expected
levels of accuracy. In Chapter 21 it is shown that when the control is included
as observations, poor observations and control stations can be isolated in the
adjustment by using weights.
Example 20.1 A trilateration survey was completed for the network shown
in Figure 20.1 and the following observations collected:
Figure 20.1 Trilateration network.
390
CONSTRAINT EQUATIONS
Control stations
Station
X (ft)
Y (ft)
A
C
10,000.00
12,487.08
10,000.00
10,528.65
Distance observations
From
To
Distance (ft)
(ft)
From
To
Distance (ft)
(ft)
A
A
B
C
C
B
E
C
F
D
1400.91
1090.55
1723.45
976.26
1244.40
0.023
0.022
0.023
0.022
0.023
B
B
D
D
E
E
F
F
E
F
1644.29
1217.54
842.75
1044.99
930.93
0.023
0.022
0.022
0.022
0.022
Perform a least squares adjustment of this survey, holding the control coordinates of stations A and C by appropriate weights (assume that these control stations have a precision of 1⬊10,000).
SOLUTION The J, X, and K matrices formed in this adjustment are
⭸DAB
⭸xA
⭸DAB
⭸yA
⭸DAB
⭸xB
⭸DAB
⭸yB
0
0
0
0
0
0
0
0
⭸DAE
⭸xA
⭸DAE
⭸yA
0
0
0
0
0
0
⭸DAE
⭸xE
⭸DAE
⭸yE
0
0
0
0
⭸DBE
⭸xB
⭸DBE
⭸yB
0
0
0
0
⭸DBE
⭸xE
⭸DBE
⭸yE
0
0
0
0
⭸DBF
⭸xB
⭸DBF
⭸yB
0
0
0
0
0
0
⭸DBF
⭸xF
⭸DBF
⭸yF
0
0
⭸DBC
⭸xB
⭸DBC
⭸yB
⭸DBC
⭸xC
⭸DBC
⭸yC
0
0
0
0
0
0
0
0
0
0
⭸DCF
⭸xC
⭸DCF
⭸yC
0
0
0
0
⭸DCF
⭸xF
⭸DCF
⭸yF
0
0
0
0
⭸DCD
⭸xC
⭸DCD
⭸yC
⭸DCD
⭸xD
⭸DCD
⭸yD
0
0
0
0
0
0
0
0
0
0
⭸DDF
⭸xD
⭸DDF
⭸yD
0
0
⭸DDF
⭸xF
⭸DDF
⭸yF
0
0
0
0
0
0
⭸DDE
⭸yD
⭸DDE
⭸yD
⭸DDE
⭸xE
⭸DDE
⭸yE
0
0
0
0
0
0
0
0
0
0
⭸DEF
⭸xE
⭸DEF
⭸yE
⭸DEF
⭸xF
⭸DEF
⭸yF
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
J⫽
20.2
ADJUSTMENT OF CONTROL STATION COORDINATES
391
LAB ⫺ AB0
LAE ⫺ AE0
LBE ⫺ BE0
LBF ⫺ BF0
LBC ⫺ BC0
LCF ⫺ CF0
LCD ⫺ CD0
K⫽
LDF ⫺ DF0
LDE ⫺ DE0
LEF ⫺ EF0
xA ⫺ xA0
yA ⫺ yA0
xC ⫺ xC0
yC ⫺ yC0
dxA
dyA
dxB
dyB
dxC
dyC
X⫽
dxD
dyD
dxE
dyE
dxF
dyF
Notice that the last four rows of the J matrix correspond to observation
equations (20.1) for the coordinates of control stations A and C. Each coordinate has a row with one in the column corresponding to its correction.
Obviously, by including the control station coordinates, four unknowns
have been added to the adjustment: dxA, dyA, dxC, and dyC. However, four
observations have also been added. Therefore, the number of redundancies
is unaffected by adding the coordinate observation equations. That is, the
adjustment has the same number of redundancies with or without the control
equations.
It is possible to weight a control station according to the precision of its
coordinates. Unfortunately, control stations are published with distance precisions rather than the covariance matrix elements that are required for
weighting. However, estimates of the standard deviations of the coordinates
can be computed from the published distance precisions. That is, if the distance precision between stations A and C is 1⬊10,000 or better, their coordinates should have estimated errors that yield a distance precision of
1⬊10,000 between the stations. To find the estimated errors in the coordinates
that yield the appropriate distance precision between the stations, Equation
(6.16) can be applied to the distance formula, resulting in
2
⭸Dij
⭸xi
2
D
⫽
ij
x2i ⫹
⭸Dij
⭸yi
2
y2i ⫹
⭸Dij
⭸xj
2
x2j ⫹
⭸Dij
⭸yj
2
y2j
(20.2)
In Equation (20.2), 2Dij is the variance in distance Dij, and x2i, y2i, x2j, and
2yj are the variances in the coordinates of the endpoints of the line. Assuming
that the estimated errors for the coordinates of Equation (20.2) are equal and
substituting in the appropriate partial derivatives yields
2Dij ⫽ 2
⭸Dij
⭸x
2
2x ⫹ 2
⭸Dij
⭸y
2
2y ⫽ 2
x
IJ
2
c
⫹2
x
IJ
2
c
(20.3)
392
CONSTRAINT EQUATIONS
In Equation (20.3), c is the standard deviation in the x and y coordinates.
(Note that the partial derivatives appearing in Equation (20.3) were described
in Section 14.2.) Factoring 22c from Equation (20.3) yields
x2 ⫹ y2
2Dij ⫽ 22c
⫽ 2c2
IJ2
(20.4)
where
2c ⫽ 2x ⫽ 2y .
From the coordinates of A and C, distance AC is 2542.65 ft. To get a
distance precision of 1⬊10,000, a maximum distance error of 0.25 ft exists.
Assuming equal coordinate errors,
0.25 ⫽ x 兹2 ⫽ y 兹2
Thus, x ⫽ y ⫽ 0.18 ft. The standard deviations computed are used to
weight the control in the adjustment. The weight matrix for this adjustment
is
W⫽
1
0.0232
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0222
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0232
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0222
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0232
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0222
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0222
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0222
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0222
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0.0222
1
0
0
0
0.182
1
0
0
0
0.182
1
0
0
0
0.182
0
0
0
0
0
0
0
0
1
0.182
20.2
ADJUSTMENT OF CONTROL STATION COORDINATES
393
The adjustment results, obtained using the program ADJUST, are shown
below.
**** Adjusted Distance Observations ****
No. 兩
From 兩
To 兩
Distance 兩
Residual 兩
========================================================
1兩
A 兩
B 兩
1,400.910 兩
⫺0.000 兩
2兩
A 兩
E 兩
1,090.550 兩
⫺0.000 兩
3兩
B 兩
E 兩
1,644.288 兩
⫺0.002 兩
4兩
B 兩
F 兩
1,217.544 兩
0.004 兩
5兩
B 兩
C 兩
1,723.447 兩
⫺0.003 兩
6兩
C 兩
F 兩
976.263 兩
0.003 兩
7兩
C 兩
D 兩
1,244.397 兩
⫺0.003 兩
8兩
D 兩
E 兩
1,044.988 兩
⫺0.002 兩
9兩
E 兩
F 兩
930.933 兩
0.003 兩
10兩
D 兩
F 兩
842.753 兩
0.003 兩
=================
****** Adjusted Control Stations ******
No. 兩 Sta. 兩
Northing 兩
Easting 兩
N Res 兩
E Res 兩
========================================================
1兩
C 兩 10,528.650 兩 12,487.080 兩
0.000 兩
0.002 兩
2兩
A 兩 10,000.000 兩 10,000.000 兩 ⫺0.000 兩 ⫺0.002 兩
========================================================
Reference Standard Deviation ⫽ 0.25
Degrees of Freedom ⫽ 2
******* Adjusted Unknowns ******
Station兩 Northing 兩
Easting 兩 North 兩 East 兩 t ang⬚ 兩 A axis 兩 B axis 兩
======================================================================
A兩 10,000.000兩 9,999.998兩
0.033兩
0.045兩 168.000兩
0.046兩
0.033兩
B兩 11,103.933兩 10,862.483兩
0.039兩
0.034兩 65.522兩
0.040兩
0.033兩
C兩 10,528.650兩 12,487.082兩
0.033兩
0.045兩 168.000兩
0.046兩
0.033兩
D兩 9,387.462兩 11,990.882兩
0.040兩
0.038兩 49.910兩
0.044兩
0.033兩
E兩 9,461.900兩 10,948.549兩
0.039兩
0.034兩 110.409兩
0.039兩
0.033兩
F兩 10,131.563兩 11,595.223兩
0.033兩
0.034兩 17.967兩
0.034兩
0.033兩
======================================================================
Notice that the control stations were adjusted slightly, as evidenced by their
residuals. Also note the error ellipse data computed for each control station.
394
CONSTRAINT EQUATIONS
20.3 HOLDING CONTROL STATION COORDINATES AND
DIRECTIONS OF LINES FIXED IN A TRILATERATION ADJUSTMENT
As demonstrated in Example 14.1, the coordinates of a control station are
easily fixed during an adjustment. This is accomplished by assigning values
of zero to the coefficients of the dx and dy correction terms. This method
removes their corrections from the equations. In that particular example, each
observation equation had only two unknowns, since one end of each observed
distance was a control station that was held fixed during the adjustment.
This was a special case of a method known as solution by elimination of
constraints.
This method can be shown in matrix notation as
A1X1 ⫹ A2X2 ⫽ L1 ⫹ V
(20.5)
C1X1 ⫹ C2X2 ⫽ L2
(20.6)
In Equation (20.6), A1, A2, X1, X2, L1, and L2 are the A, X, and L matrices
partitioned by the constraint equations, as shown in Figure 20.2; C1 and C2
are the partitions of the matrix C, consisting of the coefficients of the constraint equations; and V is the residual matrix. In this method, matrices A, C,
and X are partitioned into two matrix equations that separate the constrained
and unconstrained observations. Careful consideration should be given to the
partition of C1 since this matrix cannot be singular. If singularity exists, a
new set of constraint equations that are mathematically independent must be
determined. Also, since each constraint equation will remove one parameter
from the adjustment, the number of constraints must not be so large that the
remaining A1 and X1 have no independent equations or are themselves
singular.
From Equation (20.6), solve for X1 in terms of C1, C2, X2, and L2 as
X1 ⫽ C1⫺1(L2 ⫺ C2X2)
Substituting Equation (20.7) into Equation (20.5) yields
Figure 20.2 A, X, and L matrices partitioned by constraint equations.
(20.7)
20.3
HOLDING CONTROL STATION COORDINATES AND DIRECTIONS OF LINES FIXED
A1 [C1⫺1(L2 ⫺ C2X2)] ⫹ A2X2 ⫽ L1 ⫹ V
395
(20.8)
Rearranging Equation (20.8), regrouping, and dropping V for the time being
gives
(⫺A1C1⫺1C2 ⫹ A2)X2 ⫽ L1 ⫺ A1C1⫺1L2
(20.9)
Letting A⬘ ⫽ ⫺A1C⫺1
1 C2 ⫹ A2, Equation (20.9) can be rewritten as
A⬘X2 ⫽ L1 ⫺ A1C1⫺1L2
(20.10)
Now Equation (20.10) can be solved for X2, which, in turn, is substituted into
Equation (20.7) to solve for X1.
It can be seen that in the solution by elimination of constraint, the constraints equations are used to eliminate unknown parameters from the adjustment, thereby fixing certain geometric conditions during the adjustment. This
method was used when the coordinates of the control stations were removed
from the adjustments in previous chapters. In the following subsection, this
method is used to hold the azimuth of a line during an adjustment.
20.3.1 Holding the Direction of a Line Fixed by Elimination
of Constraints
Using this method, constraint equations are written and then functionally substituted into the observation equations to eliminate unknown parameters. To
illustrate, consider that in Figure 20.3, the direction of a line IJ is to be held
fixed during the adjustment. Thus, the position of J is constrained to move
linearly along IJ during the adjustment. If J moves to J⬘ after adjustment, the
relationship between the direction of IJ and dxj and dyj is
dxj ⫽ dyj tan ␣
(20.11)
For example, suppose that the direction of line AB in Figure 20.4 is to be
held fixed during a trilateration adjustment. Noting that station A is to be held
Figure 20.3 Holding direction IJ fixed.
396
CONSTRAINT EQUATIONS
Figure 20.4 Holding direction AB fixed in a trilateration adjustment.
fixed and using prototype equation (14.9), the following observation equation
results for observed distance AB:
klab ⫹ vlb ⫽
xb0 ⫺ xa
yb ⫺ ya
dxb ⫹ 0
dyb
AB0
AB0
(20.12)
Now based on Equation (20.11), the following relationship is written for line
AB:
dxb ⫽ dyb tan ␣
(20.13)
Substituting Equation (20.13) into Equation (20.12) yields
klab ⫹ vlab ⫽
xb0 ⫺ xa
yb ⫺ ya
tan ␣ dyb ⫹ 0
dyb
AB0
AB0
(20.14)
Factoring dyb in Equation (20.14), the constrained observation equation is
klab ⫹ vlab ⫽
(xb0 ⫺ xa) tan ␣ ⫹ (yb0 ⫺ ya)
dyb
AB0
(20.15)
Using this same method, the coefficients of dyb for lines BC and BD are
also determined, resulting in the J matrix shown in Table 20.1.
For this example, the K, X, and V matrices are
冤 冥 冤冥 冤冥
AB ⫺ AB0
AC ⫺ AC0
AD ⫺ AD0
K⫽
BC ⫺ BC0
BD ⫺ BD0
CD ⫺ CD0
dyb
dyc
X ⫽ dxc
dyd
dxd
V⫽
vab
vac
vad
vbb
vbd
vcd
TABLE 20.1 J Matrix of Figure 20.3
Unknown
Distance
dyb
dyc
dxc
dxd
dyd
AB
AC
AD
BC
BD
CD
[(xb ⫺ xa) tan ␣ ⫹ (yb ⫺ ya)] / AB
0
0
[(xb ⫺ xc) tan ␣ ⫹ (yb ⫺ yc)] / BC
[(xb ⫺ xd) tan ␣ ⫹ (yb ⫺ yd)] / BD
0
0
(yc ⫺ ya) / AC
0
(yc ⫺ yb) / BC
0
(yc ⫺ yd) / CD
0
(xc ⫺ xa) / AC
0
(xc ⫺ xb) / BC
0
(xc ⫺ xd) / CD
0
0
(yd ⫺ ya) / AD
0
(yd ⫺ yb) / BD
(yd ⫺ yc) / CD
0
0
(xd ⫺ xa) / AD
0
(xd ⫺ xb) / BD
(xd ⫺ xc) / CD
397
398
20.4
CONSTRAINT EQUATIONS
HELMERT’S METHOD
Another method of introducing constraints was presented by F. R. Helmert in
1872. In this procedure, the constraint equation(s) border the reduced normal
equations as
冤
ATWA
C
⯗
⯗
CT
冥冤 冥 冤 冥
ATWL1
X1
0
X2
⫽
(20.16)
L2
To establish this matrix, the normal matrix and its matching constants matrix
are formed, as has been done in Chapters 13 through 19. Following this, the
observation equations for the constraints are formed. These observation equations are then included in the normal matrix as additional rows [C] and columns [C T] in Equation (20.16) and their constants are added to the constants
matrix as additional rows [L2] in Equation (20.16). The inverse of this
bordered normal matrix is computed. The matrix solution of the Equation
(20.16) is
冤 冥 冤
ATWA
X1
X2
⫽
C
⯗
⯗
CT
冥冤 冥
⫺1
ATWL1
0
L2
(20.17)
In Equation (20.17) X2 is not used in the subsequent solution for the unknowns. This procedure is illustrated in the following examples.
Example 20.2 Constrained Differential Leveling Adjustment In Figure
20.5, differential elevations were observed for a network where the elevation
difference between stations B and E is to be held at ⫺17.60 ft. The elevation
of A is 1300.62 ft, and the elevation differences observed for each line are
shown below.
Figure 20.5 Differential leveling network.
20.4
Line
From
To
1
2
3
4
5
6
7
A
B
C
D
C
E
E
B
C
D
A
E
D
A
HELMERT’S METHOD
Elevation (ft)
399
S (ft)
25.15
0.07
0.05
0.03
0.08
0.03
0.05
0.05
⫺10.57
⫺1.76
⫺12.65
⫺7.06
5.37
⫺7.47
Perform a least squares adjustment of this level net constraining the required
elevation difference.
SOLUTION The A, X, and L matrices are
冤
1
⫺1
A⫽
0
0
0
0
0
冥
0
0
0
1
0
0
⫺1
1
0
0 ⫺1
0
⫺1
0
1
0
1 ⫺1
0
0 ⫺1
冤 冥
1325.77
⫺10.55
⫺1.76
L ⫽ ⫺1313.27
⫺7.06
5.37
1308.09
冤冥
B
C
X⫽
D
E
The weight matrix (W) is
冤
冥
204.08
0
0
0
0
0
0
0 400
0
0
0
0
0
0
0 1111.11
0
0
0
0
W⫽
0
0
0 156.25
0
0
0
0
0
0
0 1111.11
0
0
0
0
0
0
0 400
0
0
0
0
0
0 400
The reduced normal equations are
冤
604.08
冥冤 冥 冤
⫺400.00
0.00
0.00
2622.22 ⫺1111.11 ⫺1111.11
0.00 ⫺1111.11
1667.36
⫺400.00
0.00 ⫺1111.11 ⫺400.00
1911.11
⫺400.00
冥
B
274,793.30
C
5572.00
⫽
D
205,390.90
E
513,243.60
The reduced normal matrix is now bordered by the constraint equation
E ⫺ B ⫽ ⫺17.60
which has a matrix form of
(a)
400
CONSTRAINT EQUATIONS
[⫺1
0
0
冤冥
B
C
1]
⫽ [⫺17.60]
D
E
(b)
The left side of Equation (b) is now included as an additional row and column
to the reduced normal matrix in Equation (a). The lower-right corner diagonal
element of the newly bordered normal matrix is assigned a value of 0. Similarly, the right-hand side of Equation (b) is added as an additional row in the
right-hand side of Equation (a). Thus, the bordered-normal equations are
冤
604.80
⫺400.00
0.00
0.00 ⫺1
2622.22 ⫺1111.11 ⫺1111.11
0
0.00 ⫺1111.11
1667.36
⫺400.00
0
0.00 ⫺1111.11 ⫺400.00
1911.11
1
⫺1
0
0
1
0
⫺400.00
冥冤 冥 冤 冥
B
274,793.30
C
5572.00
D ⫽ 205,390.90
E
513,243.60
X2
⫺17.60
(c)
Notice in Equation (c) that an additional unknown, X2, is added at the bottom
of the X to make the X matrix dimensionally consistent with the borderednormal matrix of (c). Similarly, on the right-hand side of the equation, the k
value of constraint equation (b) is added to the bottom of the matrix. Using
Equation (20.17), the resulting solution is
冤 冥
1325.686
1315.143
X ⫽ 1313.390
1308.086
⫺28.003
(d)
From the X matrix in Equation (d), elevation of station B is 1325.686 and
that for station E is 1308.086. Thus, the elevation difference between stations
B and E is exactly ⫺17.60, which was required by the constraint condition.
Example 20.3 Constraining the Azimuth of a Line Helmert’s method
can also be used to constrain the direction of a line. In Figure 20.6 the bearing
of line AB is to remain at its record value of N 0⬚04⬘ E. The data for this
trilaterated network are
Control station
Initial approximations
Station
X (m)
Y (m)
Station
X (m)
Y (m)
A
1000.000
1000.000
B
C
D
1003.07
2323.07
2496.08
3640.00
3638.46
1061.74
20.4
HELMERT’S METHOD
401
Figure 20.6 Network for Example 20.3.
Distance observations
From
To
Distance (m)
(m)
From
To
Distance (m)
(m)
A
A
B
C
B
C
2951.604
2640.017
1320.016
0.025
0.024
0.021
C
D
B
D
A
D
2582.534
1497.360
2979.325
0.024
0.021
0.025
Adjust the figure by the method of least squares holding the direction of the
line AB using Helmert’s method.
SOLUTION Using procedures discussed in Chapter 14, the reduced normal
equations for the trilaterated system are
冤
2690.728
⫺706.157
⫺706.157
⫺2284.890
2988.234
0.000
0.000
0.000
405.837
706.157
706.157 ⫺1228.714
冥
⫺2284.890
0.000 ⫺405.837
706.157
0.000
0.000
706.157 ⫺1228.714
2624.565
529.707
⫺8.737
124.814
529.707
3077.557
124.814 ⫺1783.060
⫺8.737
124.814 2636.054 ⫺742.112
124.814 ⫺1783.060 ⫺742.112
3015.328
冤冥冤 冥
dxb
⫺6.615
dyb
6.944
dxc
21.601
⫻
⫽
dyc
17.831
dxd
7.229
dyd
11.304
(e)
Following prototype equation (15.9), the linearized equation for the azimuth
of line AB is
[78.13 ⫺0.09
0
0
0
冤冥
dxb
dyb
dxc
0]
⫽ [0.139]
dyc
dxd
dyd
(ƒ)
402
CONSTRAINT EQUATIONS
The observation equation for the constrained direction [Equation (ƒ)] is then
added to the border of the matrix of reduced normal equations (e), which
yields
冤
冥
2690.728 ⫺706.157 ⫺2284.890
0.000 ⫺405.837
706.157 78.13
2988.234
0.000
0.000 706.157 ⫺1228.714 ⫺0.09
0.000 2624.565
529.707 ⫺8.737
124.814 0.000
0.000
0.000
529.707 3077.557 124.814 ⫺1783.060 0.000
405.837
706.157
⫺8.737
124.814 2636.054 ⫺742.112 0.000
706.157 ⫺1228.714
124.814 ⫺1783.060 ⫺742.112 3015.328 0.000
78.13
⫺0.09
0.000
0.000
0.000
0.000 0.000
⫺706.157
⫺2284.890
冤冥冤 冥
dxb
⫺6.615
dyb
6.944
dxc
21.601
⫻ dyc ⫽ 17.831
dxd
7.229
dyd
11.304
dx2
0.139
(g)
This is a nonlinear problem, and thus the solution must be iterated until
convergence. The first two iterations yielded the X matrices listed as X1 and
X2 below. The third iteration resulted in negligible corrections to the unknowns. The total of these corrections in shown below as XT.
冤 冥 冤 冥 冤 冥
0.00179
0.00799
0.00636
X1 ⫽
0.01343
0.00460
0.01540
⫺0.00337
⫺0.00001
⫺0.00553
0.00477
X2 ⫽ ⫺0.00508
⫺0.00342
⫺0.00719
⫺0.00359
0.00178
0.00247
0.01113
XT ⫽
0.00835
0.00117
0.00821
⫺0.00696
Adding the coordinate corrections of XT to the initial approximations results
in the final coordinates for stations B, C, and D of
B: (1003.072, 3640.003) C: (2323.081, 3638.468) D: (2496.081, 1061.748)
Checking the solution: Using Equation (15.1), check to see that the direction of line AB was held to the value of the constraint.
AzAB ⫽ tan⫺1
3.072
⫽ 0⬚04⬘00ⴖ (Check!)
2640.003
20.6
20.5
ENFORCING CONSTRAINTS THROUGH WEIGHTING
403
REDUNDANCIES IN A CONSTRAINED ADJUSTMENT
The number of redundancies in an adjustment increases by one for each parameter that is removed by a constraint equation. An expression for determining the number of redundancies is
r⫽m⫺n⫹c
(20.18)
where r is the number of redundancies (degrees of freedom) in the system,
m the number of observations in the system, n the number of unknown parameters in the system, and c the number of mathematically independent
constraints applied to the system. In Example 20.2 there were seven observations in a differential leveling network that had four stations with unknown
elevations. One constraint was added to the system of equations that fixed the
elevation difference between B and E as ⫺17.60. In this way, the elevation
of B and E became mathematically dependent. By applying Equation (20.18),
it can be seen that the number of redundancies in the system is r ⫽ 7 ⫺ 4
⫹ 1 ⫽ 4. Without the aforementioned constraint, this adjustment would have
only 7 ⫺ 4 ⫽ 3 redundancies. Thus, the constraint added one degree of
freedom to the adjustment while making the elevations of B and E mathematically dependent.
Care must be used when adding constraints to an adjustment. It is possible
to add as many mathematically independent constraint equations as there are
unknown parameters. If that is done, all unknowns are constrained or fixed,
and it is impossible to perform an adjustment. Furthermore, it is also possible
to add constraints that are mathematically dependent equations. Under these
circumstances, even if the system of equations has a solution, two mathematically dependent constraints would remove only one unknown parameter,
and thus the redundancies in the system would increase by only one.
20.6
ENFORCING CONSTRAINTS THROUGH WEIGHTING
The methods described above for handling constraint equations can often be
avoided simply by overweighting the observations to be constrained in a
weighted least squares adjustment. This was done in Example 16.2 to fix the
direction of a line. As a further demonstration of the procedure of enforcing
constraints by overweighting, Example 20.3 will be adjusted by writing observation equations for azimuth AB and the control station coordinates XA and
YA. These observations will be fixed by assigning a 0.001ⴖ standard deviation
to the azimuth of line AB and standard deviations of 0.001 ft to the coordinates of station A.
404
CONSTRAINT EQUATIONS
The J, K, and W matrices for the first iteration of this problem are listed
below. Note that the numbers have been rounded to three-decimal places for
publication purposes only.
J⫽
⫺0.448
⫺0.001
⫺0.894
⫺1.000
0.000
0.000
0.000
0.000
⫺0.999 ⫺0.041
0.000
0.000
⫺78.130
0.091
1.000
0.000
0.000
1.000
0.000
0.000
0.448
0.894
0.001
1.000
0.000
0.000
⫺1.000
0.001
1.000 ⫺0.001
0.000
0.000 ⫺0.067
0.998
0.000
0.000
0.000
0.000
⫺0.050
0.865
0.000
0.000
78.130 ⫺0.091
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.000
0.067 ⫺0.997
0.999
0.041
0.050 ⫺0.865
0.000
0.000
0.000
0.000
0.000
0.000
⫺0.003
0.015
0.015
0.012
K ⫽ 0.007
⫺0.021
0.139
0.000
0.000
1
0.0252
0
W⫽
0
0
1
0.0242
0
0
0
0
0
0
0
0
0
0
0
1
0.0212
0
0
0
0
0
0
0
0
0
0
0
1
0.0242
0
0
0
0
0
0
0
1
0.0212
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0.0012
0
0
0
0
0
0
0
0
1
0.0252
0
0
0
0
0
0
0
0.0012
0
0
0
0
0
0
0
0
2
0
1
0.0012
The results of the adjustment (from the program ADJUST) are presented
below.
20.6
ENFORCING CONSTRAINTS THROUGH WEIGHTING
405
*****************
Adjusted stations
*****************
Standard error ellipses computed
Station
X
Y
Sx
Sy
Su
Sv
t
==============================================================
A
1,000.000 1,000.000 0.0010 0.0010 0.0010 0.0010 135.00⬚
B
1,003.072 3,640.003 0.0010 0.0217 0.0217 0.0010
0.07⬚
C
2,323.081 3,638.468 0.0205 0.0248 0.0263 0.0186 152.10⬚
D
2,496.081 1,061.748 0.0204 0.0275 0.0281 0.0196 16.37⬚
*******************************
Adjusted Distance Observations
*******************************
Station
Station
Occupied
Sighted
Distance
V
S
========================================================
A
C
2,951.620
0.0157
0.0215
A
B
2,640.004
⫺0.0127
0.0217
B
C
1,320.010
⫺0.0056
0.0205
C
D
2,582.521
⫺0.0130
0.0215
D
A
1,497.355
⫺0.0050
0.0206
B
D
2,979.341
0.0159
0.0214
*****************************
Adjusted Azimuth Observations
*****************************
Station
Station
Occupied
Sighted
Azimuth
V
Sⴖ
========================================================
A
B
0⬚ 04⬘ 00ⴖ
0.0ⴖ
0.0ⴖ
****************************************
Adjustment Statistics
****************************************
Iterations
Redundancies
Reference Variance
Reference So
⫽
⫽
⫽
⫽
2
1
1.499
1.2
Passed X2 test at 95.0% significance level!
X2 lower value ⫽ 0.00
406
CONSTRAINT EQUATIONS
X2 upper value ⫽ 5.02
A priori value of 1 used for reference variance
in computations of statistics.
Convergence!
Notice in the adjustment above that the control station coordinates remained fixed and the residual of the azimuth of line AB is zero. Thus, the
azimuth of line AB was held fixed without the inclusion of any constraint
equation. It was simply constrained by overweighting the observation. Also
note that the final adjusted coordinates of stations B, C, and D match the
solution in Example 20.3.
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
20.1
Given the following lengths observed in a trilateration survey, adjust
the survey by least squares using the elimination of constraints
method to hold the coordinates of A at xa ⫽ 30,000.00 and ya ⫽
30,000.00, and the azimuth of line AB to 30⬚00⬘00.00ⴖ 0.001ⴖ from
north. Find the adjusted coordinates of B, C, and D.
Distance observations
Course
Distance (ft)
S (ft)
Course
Distance (ft)
S (ft)
AB
BC
CD
22,867.12
22,943.74
28,218.26
0.116
0.116
0.142
DA
AC
BD
29,593.60
30,728.64
41,470.07
0.149
0.155
0.208
Initial coordinates
Station
X (ft)
Y (ft)
B
C
D
41,433.56
60,054.84
51,386.93
49,803.51
36,399.65
9,545.64
20.2
Do Problem 20.1 using Helmert’s method.
20.3
For the following traverse data, use Helmert’s method and perform
an adjustment holding the coordinates of station A fixed and azimuth
of line AB fixed at 269⬚28⬘11ⴖ. Assume that all linear units are feet.
PROBLEMS
Control station
Initial coordinates
Station
X
Y
Station
X
Y
A
15,123.65
9803.10
B
C
14,423.26
12,620.56
9796.61
9066.30
Distance observations
20.4
407
Angle observations
Course
Distance
S
Stations
Angle
S (ⴖ)
AB
BC
CA
700.42
1945.01
2609.24
0.020
0.022
0.024
ABC
BCA
CAB
158⬚28⬘34ⴖ
5⬚39⬘06ⴖ
15⬚52⬘22ⴖ
4.5
2.9
4.1
Given the following differential leveling data, adjust it using Helmert’s method. Hold the elevation of A to 100.00 ft and the elevation
difference ElevAD to ⫺5.00 ft.
From
To
Elev (ft)
A
B
C
B
C
D
⫺19.984
⫺4.998
19.997
S
From
To
Elev (ft)
0.005
0.008
0.003
D
E
E
A
⫺5.000
9.990
S
0.006
0.004
20.5
Using the method of weighting discussed in Section 20.6, adjust the
data in:
(a) Problem 20.1.
(b) Problem 20.3.
(c) Problem 20.4.
(d) Compare the results with those of Problem 20.1, 20.3, or 20.4 as
appropriate.
20.6
Do
(a)
(b)
(c)
20.7
Do Problem 14.15 holding the elevation difference between V and Z
to 3.600 m using:
(a) the elimination of constraints method.
(b) Helmert’s method.
(c) the method of overweighting technique.
(d) Compare the results of the adjustments from the various methods.
Problem 13.14 holding distance BC to 100.00 ft using:
the elimination of constraints method.
Helmert’s method.
Compare the results of the adjustments from the different
methods.
408
CONSTRAINT EQUATIONS
20.8
Do Problem 12.11 holding the difference in elevation between stations
2 and 8 to 66.00 ft. Use:
(a) the elimination of constraints method.
(b) Helmert’s method.
(c) the method of overweighting technique.
20.9
Assuming that stations A and D are second-order class I horizontal
control (1⬊50,000), do Problem 14.8 by including the control in the
adjustment.
20.10 Do Problem 15.7 assuming that station A is first-order horizontal control (1⬊100,000) and B is second-order class I horizontal control
(1⬊50,000) using the overweighting method.
20.11 Do Problem 15.9 assuming that the control stations are second-order
class II horizontal control (1⬊20,000) using the overweighting
method.
20.12 Do Problem 15.12 assuming that stations A and D are third-order
class I control (1⬊10,000) using the overweighting method.
Practical Problems
20.13 Develop a computational program that computes the coefficients for
the J matrix in a trilateration adjustment with a constrained azimuth.
Use the program to solve Problem 20.8.
20.14 Develop a computational program that computes a constrained least
squares adjustment of a trilateration network using Helmert’s method.
Use this program to solve Problem 20.11.
CHAPTER 21
BLUNDER DETECTION IN
HORIZONTAL NETWORKS
21.1
INTRODUCTION
Up to this point, data sets are assumed to be free of blunders. However, when
adjusting real observations, the data sets are seldom blunder free. Not all
blunders are large, but no matter their sizes, it is desirable to remove them
from the data set. In this chapter, methods used to detect blunders before and
after an adjustment are discussed.
Many examples can be cited that illustrate mishaps that have resulted from
undetected blunders in survey data. However, few could have been more
costly and embarrassing than a blunder of about 1 mile that occurred in an
early nineteenth-century survey of the border between the United States and
Canada near the north end of Lake Champlain. Following the survey, construction of a U.S. military fort was begun. The project was abandoned two
years later when the blunder was detected and a resurvey showed that the
fort was actually located on Canadian soil. The abandoned facility was subsequently named Fort Blunder!
As discussed in previous chapters, observations are normally distributed.
This means that occasionally, large random errors will occur. However, in
accordance with theory, this should seldom happen. Thus, large errors in data
sets are more likely to be blunders than random errors. Common blunders in
data sets include number transposition, entry and recording errors, station
misidentifications, and others. When blunders are present in a data set, a least
squares adjustment may not be possible or will, at a minimum, produce poor
or invalid results. To be safe, the results of an adjustment should never be
accepted without an analysis of the post-adjustment statistics.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
409
410
BLUNDER DETECTION IN HORIZONTAL NETWORKS
21.2 A PRIORI METHODS FOR DETECTING BLUNDERS
IN OBSERVATIONS
In performing adjustments, it should always be assumed that there are possible
observational blunders in the data. Thus, appropriate methods should be used
to isolate and remove them. It is especially important to eliminate blunders
when the adjustment is nonlinear because they can cause the solution to diverge. In this section, several methods are discussed that can be used to isolate
blunders in a horizontal adjustment.
21.2.1
Use of the K Matrix
In horizontal surveys, the easiest method available for detecting blunders is
to use the redundant observations. When initial approximations for station
coordinates are computed using standard surveying methods, they should be
close to their final adjusted values. Thus, the difference between observations
computed from these initial approximations and their observed values (K matrix) are expected to be small in size. If an observational blunder is present,
there are two possible situations that can occur with regard to the K-matrix
values. If the observation containing a blunder is not used to compute initial
coordinates, its corresponding K-matrix value will be relatively large. However, if an observation with a blunder is used in the computation of the initial
station coordinates, the remaining redundant observations to that station
should have relatively large values.
Figure 21.1 shows the two possible situations. In Figure 21.1(a), a distance
blunder is present in line BP and is shown by the length PP⬘. However, this
distance was not used in computing the coordinates of station P, and thus the
K-matrix value for BP⬘ ⫺ BP0 will suggest the presence of a blunder by its
relatively large size. In Figure 21.1(b), the distance blunder in BP was used
to compute the initial coordinates of station P⬘. In this case, the redundant
angle and distance observations connecting P with A, C, and D may show
Figure 21.1 Presence of a distance blunder in computations.
21.2
A PRIORI METHODS FOR DETECTING BLUNDERS IN OBSERVATIONS
411
large discrepancies in the K-matrix. In the latter case, it is possible that some
redundant observations may agree reasonably with their computed values
since a shift in a station’s position can occur along a sight line for an angle
or along a radius for a distance. Still, most redundant observations will have
large K-matrix values and thus raise suspicions that a blunder exists in one
of the observations used to compute the coordinates of station P.
21.2.2
Traverse Closure Checks
As mentioned in Chapter 8, errors can be propagated throughout a traverse
to determine the anticipated closure. Large complex networks can be broken
into smaller link and loop traverses to check estimated closures against their
actual values. When a loop fails to meet its estimated closure, the observations
included in the computations should be checked for blunders.
Figure 21.2(a) and (b) show a graphical technique to isolate a traverse
distance blunder and an angular blunder, respectively. In Figure 21.2(a), a
blunder in distance CD is shown. Notice that the remaining courses, DE and
EA, are translated by the blunder in the direction of course CD. Thus, the
length of closure line (A⬘A) will be nearly equal to the length of the blunder
in CD with a direction that is consistent with the azimuth of CD. Since other
observations contain small random errors, the length and direction of the
closure line, A⬘A, will not match the blunder exactly. However, when one
blunder is present in a traverse, the misclosure and the blunder will be close
in both length and direction.
In the traverse of Figure 21.2(b), the effect of an angular blunder at traverse
station D is illustrated. As shown, the courses DE, EF, and FA⬘ will be rotated
about station D. Thus, the perpendicular bisector of the closure line AA⬘ will
point to station D. Again, due to small random errors in other observations,
the perpendicular bisector may not intersect the blunder precisely, but it
should be close enough to identify the angle with the blunder. Since the angle
at the initial station is not used in traverse computations, it is possible to
Figure 21.2 Effects of a single blunder on traverse closure.
412
BLUNDER DETECTION IN HORIZONTAL NETWORKS
isolate a single angular blunder by beginning traverse computations at the
station with the suspected blunder. In this case, when the blunder is not used
in the computations, estimated misclosure errors (see Chapter 8) will be met
and the blunder can be isolated to the single unused angle. Thus, in Figure
21.2(b), if the traverse computations were started at station D and used an
assumed azimuth for the course of CD, the traverse misclosure when returning
to D would be within estimated tolerance since the angle at D is not used in
the traverse computations.
21.3
A POSTERIORI BLUNDER DETECTION
When doing a least squares adjustment involving more than the minimum
amount of control, both a minimally and fully constrained adjustment should
be performed. In a minimally constrained adjustment, the data need to satisfy
the appropriate geometric closures and are not influenced by control errors.
After the adjustment, a 2 test1 can be used to check the a priori value of the
reference variance against its a posteriori estimate. However, this test is not
a good indicator of the presence of a blunder since it is sensitive to poor
relative weighting. Thus, the a posteriori residuals should also be checked for
the presence of large discrepancies. If no large discrepancies are present, the
observational weights should be altered and the adjustment rerun. Since this
test is sensitive to weights, the procedures described in Chapters 7 through
10 should be used for building the stochastic model of the adjustment.
Besides the sizes of the residuals, the signs of the residuals may also
indicate a problem in the data. From normal probability theory, residuals are
expected to be small and randomly distributed. A small section of a larger
network is shown in Figure 21.3. Notice that the distance residuals between
stations A and B are all positive. This is not expected from normally distributed data. Thus, it is possible that either a blunder or a systematic error is
present in some or all of the survey. If both A and B are control stations, part
of the problem could stem from control coordinate discrepancies. This possibility can be isolated by doing a minimally constrained adjustment.
Although residual sizes can suggest observational errors, they do not necessarily identify the observations that contain blunders. This is due to the fact
that least squares generally spreads a large observational error or blunder out
radially from its source. However, this condition is not unique to least squares
adjustments since any arbitrary adjustment method, including the compass
rule for traverse adjustment, will also spread a single observational error
throughout the entire observational set.
1
Statistical testing was discussed in Chapter 4.
21.3
A POSTERIORI BLUNDER DETECTION
413
Figure 21.3 Distribution of residuals by sign.
Although an abnormally large residual may suggest the presence of a blunder in an observation, this is not always true. One reason for this could be
poor relative weighting in the observations. For example, suppose that angle
GAH in Figure 21.4 has a small blunder but has been given a relatively high
weight. In this case the largest residual may well appear in a length between
stations G and H, B and H, C and F, and most noticeably between D and E,
due to their distances from station A. This is because the angular blunder will
cause the network to spread or compress. When this happens, the signs of
the distance residuals between G and H, B and H, C and F, and D and E
may all be the same and thus indicate the problem. Again this situation can
Figure 21.4 Survey network.
414
BLUNDER DETECTION IN HORIZONTAL NETWORKS
be minimized by using proper methods to determine observational variances
so that they truly reflect the estimated errors in the observations.
21.4 DEVELOPMENT OF THE COVARIANCE MATRIX FOR
THE RESIDUALS
In Chapter 5 it was shown how a sample data set could be tested at any
confidence level to isolate observational residuals that were too large. The
concept of statistical blunder detection in surveying was introduced in the
mid-1960s and utilizes the cofactor matrix for the residuals. To develop this
matrix, the adjustment of a linear problem can be expressed in matrix form
as
L ⫹ V ⫽ AX ⫹ C
(21.1)
where C is a constants vector, A the coefficient matrix, X the estimated parameter matrix, L the observation matrix, and V the residual vector. Equation
(21.1) can be rewritten in terms of V as
V ⫽ AX ⫺ T
(21.2)
where T ⫽ L ⫺ C, which has a covariance matrix of W⫺1 ⫽ S 2Qll. The solution
of Equation (21.2) results in the expression
X ⫽ (ATWA)⫺1 ATWT
(21.3)
Letting ε represent a vector of true errors for the observations, Equation (21.1)
can be written as
L ⫺ ε ⫽ AX ⫹ C
(21.4)
where X is the true value for the unknown parameter X and thus
T ⫽ L ⫺ C ⫽ AX ⫹ ε
(21.5)
Substituting Equations (21.3) and (21.5) into Equation (21.2) yields
V ⫽ A(ATWA)⫺1ATW(AX ⫹ ε) ⫺ (AX ⫹ ε)
(21.6)
Expanding Equation (21.6) results in
V ⫽ A(ATWA)⫺1ATWε ⫺ ε ⫹ A(ATWA)⫺1ATWAX ⫺ AX
Since (ATWA)⫺1 ⫽ A⫺1W⫺1A⫺T, Equation (21.7) can be simplified to
(21.7)
21.4
DEVELOPMENT OF THE COVARIANCE MATRIX FOR THE RESIDUALS
V ⫽ A(ATWA)⫺1ATWε ⫺ ε ⫹ AX ⫺ AX
415
(21.8)
Factoring Wε from Equation (21.8) yields
V ⫽ ⫺[W⫺1 ⫺ A(ATWA)⫺1AT]Wε
(21.9)
Recognizing (ATWA)⫺1 ⫽ Qxx and defining Qvv ⫽ W⫺1 ⫺ AQxxAT, Equation
(21.9) can be rewritten as
(21.10)
V ⫽ ⫺QvvWε
where Qvv ⫽ W⫺1 ⫺ AQxx AT ⫽ W⫺1 ⫺ Qll.
The Qvv matrix is both singular and idempotent. Being singular, it has no
inverse. When a matrix is idempotent, the following properties exist for the
matrix: (a) The square of the matrix is equal to the original matrix (i.e., Qvv
Qvv ⫽ Qvv), (b) every diagonal element is between zero and 1, and (c) the
sum of the diagonal elements, known as the trace of the matrix, equals the
degrees of freedom in the adjustment. The latter property is expressed mathematically as
q11 ⫹ q22 ⫹ 䡠 䡠 䡠 ⫹ qmm ⫽ degrees of freedom
(21.11)
(d) The sum of the square of the elements in any single row or column equals
the diagonal element. That is,
2
qii ⫽ qi1
⫹ q2i2 ⫹ 䡠 䡠 䡠 ⫹ q2im ⫽ q21i ⫹ q22i ⫹ 䡠 䡠 䡠 ⫹ q2mi
(21.12)
Now consider the case when all observations have zero errors except for
a particular observation li that contains a blunder of size ⌬li. A vector of the
true errors is expressed as
0
0
0
0
⯗
⯗
0
0
⌬ε ⫽ ⌬li εi ⫽
⫽ ⌬li
⌬li
1
0
0
⯗
⯗
0
0
(21.13)
If the original observations are uncorrelated, the specific correction for ⌬vi
can be expressed as
⌬vi ⫽ ⫺qii wii ⌬li ⫽ ⫺ri ⌬li
(21.14)
416
BLUNDER DETECTION IN HORIZONTAL NETWORKS
where qii is the ith diagonal of the Qvv matrix, wii the ith diagonal term of
the weight matrix, W, and ri ⫽ qii wii is the observational redundancy number.
When the system has a unique solution, ri will equal zero, and if the
observation is fully constrained, ri would equal 1. The redundancy numbers
provide insight into the geometric strength of the adjustment. An adjustment
that in general has low redundancy numbers will have observations that lack
sufficient checks to isolate blunders, and thus the chance for undetected blunders to exist in the observations is high. Conversely, a high overall redundancy
number enables a high level of internal checking of the observations and thus
there is a lower chance of accepting observations that contain blunders. The
quotient of r/m, where r is the total number of redundant observations in the
system and m is the number of observations, is called the relative redundancy
of the adjustment.
21.5
DETECTION OF OUTLIERS IN OBSERVATIONS
Equation (21.10) defines the covariance matrix for the vector of residuals, vi.
From this the standardized residual is computed using the appropriate diagonal element of the Qvv matrix as
vi ⫽
vi
(21.15)
兹qii
where vi is the standardized residual, vi the computed residual, and qii the
diagonal element of the Qvv matrix. Using the Qvv matrix, the standard deviation in the residual is S0兹qii. Thus, if the denominator of Equation (21.15)
is multiplied by S0, a t statistic is defined. If the residual is significantly
different from zero, the observation used to derive the statistic is considered
to be a blunder. The test statistic for this hypothesis test is
ti ⫽
vi
S0兹qii
⫽
vi
Sv
⫽
vi
S0
(21.16)
Baarda (1968) computed rejection criteria for various significance levels
(see Table 21.1) determining the ␣ and levels for Type I and Type II errors.
The interpretation of these criteria is shown in Figure 21.5. When a blunder
is present in the data set, the t distribution is shifted, and a statistical test for
this shift may be performed. As with any other statistical test, two types of
errors can occur. A Type I error occurs when data are rejected that do not
contain blunders, and a Type II error occurs when a blunder is not detected
in a data set where one is actually present. The rejection criteria are represented by the vertical line in Figure 21.5 and their corresponding significance
21.5
DETECTION OF OUTLIERS IN OBSERVATIONS
417
TABLE 21.1 Rejection Criteria with Corresponding Significance Levels
␣
1⫺
1⫺
Rejection Criterion
0.05
0.001
0.001
0.95
0.999
0.999
0.80
0.80
0.999
0.20
0.20
0.001
2.8
4.1
6.6
levels are shown in Table 21.1. In practice, authors2 have reported that 3.29
also works as a criterion for rejection of blunders.
Thus, the approach is to use a rejection level given by a t distribution with
r ⫺ 1 degrees of freedom. The observation with the largest absolute value of
ti as given by Equation (21.17) is rejected when it is greater than the rejection
level. That is, the observation is rejected when
兩vi兩
S0兹qii
⬎ rejection level
(21.17)
Since the existence of any blunder in the data set will affect the remaining
observations and since Equation (21.18) depends on S0, whose value was
computed from data containing blunders, all observations that are detected as
blunders should not be removed in a single pass. Instead, only the largest or
largest independent group of observations should be deleted. Furthermore,
since Equation (21.18) depends on S0, it is possible to rewrite the equation
so that it can be computed during the final iteration of a nonlinear adjustment.
In this case the appropriate equation is
Figure 21.5 Effects of a blunder on the t distribution.
2
References relating to the use of 3.29 as the rejection criterion are made in Amer (1979) and
Harvey (1994).
418
BLUNDER DETECTION IN HORIZONTAL NETWORKS
vi ⫽
兩vi兩
兹qii
⬎ S0 ⫻ rejection level
(21.18)
A summary of procedures for this manner of blunder detection is as
follows:
Step 1: Locate all standardized residuals that meet the rejection criteria of
Equation (21.17) or (21.18).
Step 2: Remove the largest detected blunder or unrelated blunder groups.
Step 3: Rerun the adjustment.
Step 4: Continue steps 1 through 3 until all detected blunders are removed.
Step 5: If more than one observation is removed in steps 1 through 4, reenter
the observations in the adjustment one at a time. Check the observation
after each adjustment to see if it is again detected as a blunder. If it is,
remove it from the adjustment or have that observation reobserved.
Again it should be noted that this form of blunder detection is sensitive to
improper relative weighting in observations. Thus, it is important to use
weights that are reflective of the observational errors. Proper methods of computing estimated errors in observations, and weighting, were discussed in
Chapters 7 through 10.
21.6
TECHNIQUES USED IN ADJUSTING CONTROL
As discussed in Chapter 20, some control is necessary in each adjustment.
However, since control itself is not perfect, this raises the question of how
control should be managed. If control stations that contain errors are heavily
weighted, the adjustment will improperly associate the control errors with the
observations. This effect can be removed by using only the minimum amount
of control required to fix the project. Table 21.2 lists the type of survey versus
the minimum amount of control. Thus, in a horizontal adjustment, if the
coordinates of only one station and the direction of only one line are held
TABLE 21.2 Requirements for a Minimally Constrained Adjustment
Survey Type
Differential leveling
Horizontal survey
GPS survey
Minimum Amount of Control
1
1
1
1
benchmark
point with known xy coordinates
course with known azimuth
point with known geodetic coordinates
21.6
TECHNIQUES USED IN ADJUSTING CONTROL
419
fixed, the observations will not be constricted by the control. That is, the
observations will need to satisfy the internal geometric constraints of the
network only. If more than minimum control is used, these additional constraints will be factored into the adjustment.
Statistical blunder detection can help identify weak control or the presence
of systematic errors in observations. Using a minimally constrained adjustment, the data set is screened for blunders. After becoming confident that the
blunders are removed from the data set, a fully constrained adjustment is
performed. Following the fully constrained adjustment, an F test is used to
compare the ratio of the minimally and fully constrained reference variances.
The ratio should be 1.3 If the two reference variances are found to be statistically different, two possible causes might exist. The first is that there are
errors in the control that must be isolated and removed. The second is that
the observations contain systematic error. Since systematic errors are not compensating in nature, they will appear as blunders in the fully constrained
adjustment. If systematic errors are suspected, they should be identified and
removed from the original data set and the entire adjustment procedure redone. If no systematic errors are identified,4 different combinations of control
stations should be used in the constrained adjustments until the problem is
isolated. By following this type of systematic approach, a single control station that has questionable coordinates can be isolated.
With this stated, it should be realized that the ideal amount of control in
each survey type is greater than the minimum. In fact, for all three survey
types, a minimum of three controls is always preferable. For example, in a
differential leveling survey with only two benchmarks, it would be impossible
to isolate the problem simply by removing one benchmark from the adjustment. However, if three benchmarks are used, a separate adjustment containing only two of the benchmarks can be run until the offending benchmark is
isolated.
Extreme caution should always be used when dealing with control stations.
Although it is possible that a control station was disturbed or that the original
published coordinates contained errors, that is seldom the case. A prudent
surveyor should check for physical evidence of disturbance and talk with
other surveyors before deciding to discard control. If the station was set by
a local, state, or federal agency, the surveyor should contact the proper
3
The ratio of the reference variances from the minimally and fully constrained adjustments should
be 1, since both reference variances should be statistically equal. That is, 2minimally constrained ⫽
2fully constrained.
4
When adjusting data that cover a large region (e.g., spherical excess, reduction to the ellipsoid)
it is essential that geodetic corrections to the data be considered and applied where necessary.
These corrections are systematic in nature and can cause errors when fitting to more than minimal
control.
420
BLUNDER DETECTION IN HORIZONTAL NETWORKS
authorities and report any suspected problems. People in the agency familiar
with the control may help explain any apparent problem. For example, it is
possible that the control used in the survey was established by two previously
nonconnecting surveys. In this case, the relative accuracy of the stations was
never checked when they were established. Another problem with control
common in surveys is the connection of two control points from different
datums. As an example, suppose that a first-order control station and a highaccuracy reference network (HARN) station are used as control in a survey.
These two stations come from different national adjustments and are thus in
different datums. They will probably not agree with each other in an adjustment.
21.7
DATA SET WITH BLUNDERS
Example 21.1 The network shown in Figure 21.6 was established to provide
control for mapping in the area of stations 1 through 6. It began from two
National Geodetic Survey second-order class II (1⬊20,000 precision) control
stations, 2000 and 2001. The data for the job were gathered by five field
crews in a class environment. The procedures discussed in Chapter 7 were
used to estimate the observational errors. The problem is to check for blunders
in the data set using a rejection level of 3.29S0.
Figure 21.6 Data set with blunders.
21.7
DATA SET WITH BLUNDERS
421
Control stations
Station
Northing (ft)
Easting (ft)
2000
2001
419,710.09
419,266.82
2,476,334.60
2,476,297.98
Angle observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
102
2000
102
2000
2001
201
202
203
2
3
6
1
2
2
2
2
2
2000
102
103
2001
201
202
203
3
1
5
5
5
5
3
3
3
3
2001
103
1
201
202
203
3
2
3
4
3
3
3
1
5
6
4
10910⬘54.0ⴖ
16258⬘16.0ⴖ
17201⬘43.0ⴖ
3604⬘26.2ⴖ
26354⬘18.7ⴖ
10149⬘55.0ⴖ
17649⬘10.0ⴖ
859⬘56.0ⴖ
31648⬘00.5ⴖ
32417⬘44.0ⴖ
33836⬘38.5ⴖ
26849⬘32.5ⴖ
31820⬘54.5ⴖ
5107⬘11.0ⴖ
9809⬘36.5ⴖ
7142⬘51.5ⴖ
16732⬘28.0ⴖ
25.5
28.9
11.8
7.4
9.7
8.1
8.4
6.5
6.3
8.1
10.7
9.8
7.0
7.2
10.3
15.1
14.5
Distance observations
From
To
Distance (ft)
S (ft)
2001
201
202
203
5
5
5
5
5
3
3
3
3
3
1
1
2000
102
103
201
202
203
3
3
4
6
1
2
2
1
5
6
4
2
3
102
103
1
425.90
453.10
709.78
537.18
410.46
397.89
246.61
450.67
629.58
422.70
615.74
410.44
201.98
298.10
480.71
615.74
125.24
327.37
665.79
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
0.022
422
BLUNDER DETECTION IN HORIZONTAL NETWORKS
Initial approximations were computed for the stations as follows:
Station
Northing (ft)
Easting (ft)
1
2
3
4
5
6
102
103
201
202
203
420,353.62
419,951.98
420,210.17
420,438.88
420,567.44
420,323.31
419,743.39
419,919.69
419,589.24
419,331.29
419,819.56
2,477,233.88
2,477,497.99
2,477,832.67
2,478,023.86
2,477,630.64
2,477,665.36
2,476,454.17
2,476,728.88
2,476,576.25
2,476,948.76
2,477,463.90
SOLUTION Do the a priori check of the computed observations versus their
K-matrix values. In this check, only one angle is detected as having a difference great enough to suspect that it contains a blunder. This is angle 3–5–4,
which was measured as 32417⬘44.0ⴖ but was computed as 31735⬘31.2ⴖ.
Since this difference should not create a problem with convergence during
the adjustment, the angle remained in the data set and the adjustment was
attempted. The results of the first trial adjustment are shown below. The software used the rejection criteria procedure based on Equation (21.18) for its
blunder detection. A rejection level of 3.29S0 is used for comparison against
the standardized residuals. The column headed Std. Res. represents the standardized residual of the observation as defined by Equation (21.15) and the
column headed Red. Num. represents the redundancy number of the observation as defined by Equation (21.14).
**** Adjusted Distance Observations ****
No. 兩 From 兩 To 兩 Distance 兩 Residual 兩 Std. Res. 兩 Red. Num. 兩
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
1兩
1 兩 3 兩 616.234 兩
0.494 兩
26.148 兩
0.7458 兩
2兩
1 兩 2 兩 480.943 兩
0.233 兩
12.926 兩
0.6871 兩
3兩
1 兩 3 兩 616.234 兩
0.494 兩
26.148 兩
0.7458 兩
4兩
3 兩 4 兩 267.044 兩 ⫺31.056 兩 ⫺1821.579 兩
0.6169 兩
5兩
3 兩 6 兩 203.746 兩
1.766 兩
107.428 兩
0.5748 兩
6兩
3 兩 5 兩 413.726 兩
3.286 兩
171.934 兩
0.7719 兩
7兩
3 兩 2 兩 422.765 兩
0.065 兩
3.500 兩
0.7312 兩
8兩
5 兩 2 兩 630.949 兩
1.369 兩
75.909 兩
0.6791 兩
9兩
5 兩 1 兩 449.398 兩 ⫺1.272 兩 ⫺79.651 兩
0.5377 兩
10兩
5 兩 6 兩 247.822 兩
1.212 兩
75.418 兩
0.5488 兩
11兩
5 兩 4 兩 407.125 兩
9.235 兩
631.032 兩
0.4529 兩
21.7
12兩
5
13兩 102
14兩 103
15兩 201
16兩 202
17兩 203
18兩 2000
19兩 2001
兩 3
兩 103
兩 1
兩 202
兩 203
兩 3
兩 102
兩 201
兩
兩
兩
兩
兩
兩
兩
兩
413.726
327.250
665.702
453.362
709.856
537.241
125.056
425.949
兩
兩
兩
兩
兩
兩
兩
兩
3.266
⫺0.120
⫺0.088
0.262
0.076
0.061
⫺0.184
0.049
兩
兩
兩
兩
兩
兩
兩
兩
DATA SET WITH BLUNDERS
170.888
⫺17.338
⫺12.395
91.903
10.737
8.775
⫺28.821
7.074
兩
兩
兩
兩
兩
兩
兩
兩
423
0.7719
0.1018
0.1050
0.0172
0.1048
0.1026
0.0868
0.1008
兩
兩
兩
兩
兩
兩
兩
兩
**** Adjusted Angle Observations ****
No. 兩 From 兩 Occ 兩 To 兩
Angle 兩 Residual 兩 Std. Res. 兩 Red Num 兩
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
1兩
2 兩
1 兩
3 兩 31649⬘55.1ⴖ 兩
114.6ⴖ 兩
28.041 兩 0.4164 兩
2兩
2 兩
3 兩
4 兩 16736⬘00.2ⴖ 兩
212.2ⴖ 兩
25.577 兩 0.3260 兩
3兩
2 兩
3 兩
6 兩 7143⬘01.5ⴖ 兩
10.0ⴖ 兩
1.054 兩 0.3990 兩
4兩
2 兩
3 兩
5 兩 9755⬘09.3ⴖ 兩 ⫺867.2ⴖ 兩 ⫺101.159 兩 0.6876 兩
5兩
2 兩
3 兩
1 兩 5106⬘14.6ⴖ 兩 ⫺56.4ⴖ 兩 ⫺11.156 兩 0.4985 兩
6兩 203 兩
3 兩
2 兩 859⬘36.3ⴖ 兩 ⫺19.7ⴖ 兩 ⫺13.003 兩 0.0550 兩
7兩
2 兩
5 兩
3 兩 31825⬘14.4ⴖ 兩
259.9ⴖ 兩
44.471 兩 0.6949 兩
8兩
1 兩
5 兩
3 兩 26858⬘49.8ⴖ 兩
557.3ⴖ 兩
78.590 兩 0.5288 兩
9兩
6 兩
5 兩
3 兩 33842⬘53.4ⴖ 兩
374.9ⴖ 兩
63.507 兩 0.3058 兩
10兩
3 兩
5 兩
4 兩 322ⴗ02ⴕ24.7ⴖ 兩 ⫺8119.3ⴖ 兩 ⫺1781.060 兩 0.3197 兩
11兩 2000 兩 102 兩 103 兩 16223⬘50.9ⴖ 兩 ⫺2065.1ⴖ 兩 ⫺110.371 兩 0.4194 兩
12兩 102 兩 103 兩
1 兩 17157⬘46.9ⴖ 兩 ⫺236.1ⴖ 兩 ⫺112.246 兩 0.0317 兩
13兩 2001 兩 201 兩 202 兩 26358⬘31.6ⴖ 兩
252.9ⴖ 兩
104.430 兩 0.0619 兩
14兩 201 兩 202 兩 203 兩 10152⬘56.4ⴖ 兩
181.4ⴖ 兩
57.971 兩 0.1493 兩
15兩 202 兩 203 兩
3 兩 17650⬘15.9ⴖ 兩
65.9ⴖ 兩
23.278 兩 0.1138 兩
16兩 102 兩 2000 兩 2001 兩 10940⬘18.6ⴖ 兩 1764.6ⴖ 兩
106.331 兩 0.4234 兩
17兩 2000 兩 2001 兩 201 兩 3607⬘56.4ⴖ 兩
210.2ⴖ 兩
104.450 兩 0.0731 兩
****** Adjustment Statistics ******
Adjustment’s Reference Standard Deviation ⫽ 487.79
Rejection Level ⫽ 1604.82
The proper procedure for removing blunders is to remove the single observation that is greater in magnitude than the rejection level selected for the
adjustment and is greater in magnitude than the value of any other standardized residual in the adjustment. This procedure prevents removing observations that are connected to blunders and thus are inherently affected by their
presence. By comparing the values of the standardized residuals against the
rejection level of the adjustment, it can be seen that both a single distance
(3–4) and an angle (3–5–4) are possible blunders since their standardized
residuals are greater than the rejection level chosen. However, upon inspection
of Figure 21.6, it can be seen that a blunder in distance 3–4 will directly
affect angle 3–5–4, and distance 3–4 has the standardized residual that is
424
BLUNDER DETECTION IN HORIZONTAL NETWORKS
greatest in magnitude. This explains the previous a priori rejection of this
angle observation. That is, distance 3–4 directly affects the size of angle 3–
5–4 in the adjustment. Thus, only distance 3–4 should be removed from the
observations. After removing this distance from the observations, the adjustment was rerun with the results shown below.
**** Adjusted Distance Observations ****
No. 兩 From 兩 To 兩 Distance 兩 Residual 兩 Std. Res. 兩 Red. Num. 兩
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
1兩
1 兩 3 兩 615.693 兩 ⫺0.047 兩
⫺2.495 兩
0.7457 兩
2兩
1 兩 2 兩 480.644 兩 ⫺0.066 兩
⫺3.647 兩
0.6868 兩
3兩
1 兩 3 兩 615.693 兩 ⫺0.047 兩
⫺2.495 兩
0.7457 兩
4兩 2001 兩 201 兩 425.902 兩
0.002 兩
0.265 兩
0.1009 兩
5兩
3 兩 6 兩 201.963 兩 ⫺0.017 兩
⫺1.032 兩
0.5765 兩
6兩
3 兩 5 兩 410.439 兩 ⫺0.001 兩
⫺0.032 兩
0.7661 兩
7兩
3 兩 2 兩 422.684 兩 ⫺0.016 兩
⫺0.858 兩
0.7314 兩
8兩
5 兩 2 兩 629.557 兩 ⫺0.023 兩
⫺1.280 兩
0.6784 兩
9兩
5 兩 1 兩 450.656 兩 ⫺0.014 兩
⫺0.858 兩
0.5389 兩
10兩
5 兩 6 兩 246.590 兩 ⫺0.020 兩
⫺1.241 兩
0.5519 兩
11兩
5 兩 4 兩 397.885 兩 ⫺0.005 兩
⫺0.380 兩
0.4313 兩
12兩
5 兩 3 兩 410.439 兩 ⫺0.021 兩
⫺1.082 兩
0.7661 兩
13兩 102 兩 103 兩 327.298 兩 ⫺0.072 兩 ⫺10.380 兩
0.1018 兩
14兩 103 兩 1 兩 665.751 兩 ⫺0.039 兩
⫺5.506 兩
0.1049 兩
15兩 201 兩 202 兩 453.346 兩
0.246 兩
86.073 兩
0.0172 兩
16兩 202 兩 203 兩 709.807 兩
0.027 兩
3.857 兩
0.1049 兩
17兩 203 兩 3 兩 537.193 兩
0.013 兩
1.922 兩
0.1027 兩
18兩 2000 兩 102 兩 125.101 兩 ⫺0.139 兩 ⫺21.759 兩
0.0868 兩
**** Adjusted Angle Observations ****
No. 兩From 兩 Occ 兩 To 兩
Angle
兩 Residual 兩 Std. Res.兩 Red Num兩
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
1兩
2 兩
1 兩
3 兩31647⬘54.2ⴖ 兩
⫺6.3ⴖ 兩
⫺1.551 兩 0.4160 兩
2兩
2 兩
3 兩
4 兩16732⬘31.0ⴖ 兩
3.0ⴖ 兩
0.380 兩 0.2988 兩
3兩
2 兩
3 兩
6 兩 7142⬘46.0ⴖ 兩
⫺5.5ⴖ 兩
⫺0.576 兩 0.3953 兩
4兩
2 兩
3 兩
5 兩 9809⬘18.6ⴖ 兩
⫺17.9ⴖ 兩
⫺2.088 兩 0.6839 兩
5兩
2 兩
3 兩
1 兩 5107⬘04.1ⴖ 兩
⫺6.9ⴖ 兩
⫺1.360 兩 0.4978 兩
6兩 203 兩
3 兩
2 兩 859⬘26.7ⴖ 兩
⫺29.3ⴖ 兩 ⫺19.340 兩 0.0550 兩
7兩
2 兩
5 兩
3 兩31820⬘51.4ⴖ 兩
⫺3.1ⴖ 兩
⫺0.532 兩 0.6933 兩
8兩
1 兩
5 兩
3 兩26850⬘03.4ⴖ 兩
30.9ⴖ 兩
4.353 兩 0.5282 兩
9兩
6 兩
5 兩
3 兩33836⬘37.1ⴖ 兩
⫺1.4ⴖ 兩
⫺0.238 兩 0.3049 兩
10兩
3 兩
5 兩
4 兩32417⬘43.6ⴖ 兩
⫺0.4ⴖ 兩
⫺0.381 兩 0.0160 兩
11兩2000 兩 102 兩 103 兩162ⴗ24ⴕ10.2ⴖ 兩 ⫺2045.8ⴖ 兩 ⫺109.353 兩 0.4193 兩
12兩 102 兩 103 兩
1 兩171ⴗ57ⴕ51.2ⴖ 兩 ⫺231.8ⴖ 兩 ⫺110.360 兩 0.0316 兩
13兩2001 兩 201 兩 202 兩26358⬘20.3ⴖ 兩
241.6ⴖ 兩
99.714 兩 0.0619 兩
21.7
14兩 201
15兩 202
16兩 102
17兩2000
兩 202
兩 203
兩2000
兩2001
兩 203
兩
3
兩2001
兩 201
兩10152⬘34.7ⴖ
兩17649⬘56.1ⴖ
兩109ⴗ40ⴕ17.7ⴖ
兩 3607⬘46.9ⴖ
兩
兩
兩
兩
DATA SET WITH BLUNDERS
159.7ⴖ
46.1ⴖ
1763.7ⴖ
200.7ⴖ
兩
兩
兩
兩
51.023
16.273
106.280
99.688
兩
兩
兩
兩
425
0.1494
0.1138
0.4233
0.0732
兩
兩
兩
兩
****** Adjustment Statistics ******
Adjustment’s Reference Standard Deviation ⫽ 30.62
Rejection Level ⫽ 100.73
After this adjustment, analysis of standardized residuals indicates that the
angles most likely still to contain blunders are observations 11, 12, and 16.
Of these, observation 12 displays the highest standardized residual. Looking
at Figure 21.6, it is seen that this angle attaches the northern traverse leg to
control station 2000. This is a crucial observation in the network if any hopes
of redundancy in the orientation of the network are to be maintained. Since
this is a flat angle (i.e., nearly 180), it is possible that the backsight and
foresight stations were reported incorrectly, which can be checked by reversing stations 102 and 1. However, without further field checking, it cannot be
guaranteed that this occurred. A decision must ultimately be made about
whether this angle should be reobserved. However, for now, this observation
will be discarded and another trial adjustment made. In this stepwise blunder
detection process, it is always wise to remove as few observations as possible.
In no case should observations that are blunder-free be deleted. This can and
does happen, however, in trial blunder detection adjustments. But through
persistent and careful processing, ultimately only those observations that contain blunders can be identified and eliminated. The results of the adjustment
after removing the angle 12 are shown below.
*****************
Adjusted stations
*****************
Standard error ellipses computed
Station
X
Y
Sx
Sy
Su
Sv
t
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
1 2,477,233.72 420,353.59 0.071 0.069 0.092 0.036 133.47
2 2,477,497.89 419,951.98 0.050 0.083 0.090 0.037 156.01
3 2,477,832.55 420,210.21 0.062 0.107 0.119 0.034 152.80
4 2,477,991.64 420,400.58 0.077 0.121 0.138 0.039 149.71
5 2,477,630.43 420,567.45 0.088 0.093 0.123 0.036 136.74
6 2,477,665.22 420,323.32 0.071 0.096 0.114 0.036 145.44
102 2,476,455.89 419,741.38 0.024 0.018 0.024 0.017
80.86
103 2,476,735.05 419,912.42 0.051 0.070 0.081 0.031 147.25
201 2,476,576.23 419,589.23 0.020 0.022 0.024 0.017
37.73
202 2,476,948.74 419,331.29 0.029 0.041 0.042 0.029
14.24
203 2,477,463.84 419,819.58 0.040 0.077 0.081 0.032 160.84
426
BLUNDER DETECTION IN HORIZONTAL NETWORKS
*******************************
Adjusted Distance Observations
*******************************
Station
Station
Occupied
Sighted
Distance
V
Std.Res.
Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
2001
201
425.88
⫺0.023
⫺3.25
0.102
201
202
453.09
⫺0.005
⫺3.25
0.006
202
203
709.76
⫺0.023
⫺3.25
0.104
203
3
537.16
⫺0.023
⫺3.25
0.103
5
3
410.45
⫺0.011
⫺0.60
0.767
5
4
397.89
⫺0.003
⫺0.19
0.436
5
6
246.60
⫺0.014
⫺0.83
0.556
5
1
450.68
0.013
0.80
0.542
5
2
629.58
0.003
0.15
0.678
3
2
422.70
0.003
0.16
0.736
3
1
615.75
0.008
0.40
0.745
3
5
410.45
0.009
0.44
0.767
3
6
201.97
⫺0.013
⫺0.78
0.580
1
2
480.71
⫺0.003
⫺0.19
0.688
1
3
615.75
0.008
0.40
0.745
2000
102
125.26
0.020
3.25
0.082
102
103
327.39
0.023
3.25
0.101
103
1
665.81
0.023
3.25
0.104
***************************
Adjusted Angle Observations
***************************
Station
Station
Station
Backsighted Occupied Foresighted
Angle
V Std.Res. Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
102
2000
2001 10911⬘11.1ⴖ
17.06ⴖ
3.25 0.042
2000
102
103 16258⬘05.1ⴖ ⫺10.95ⴖ
⫺3.25 0.014
2000
2001
201
3604⬘23.8ⴖ
⫺2.45ⴖ
⫺3.25 0.010
2001
201
202 26354⬘15.7ⴖ
⫺2.97ⴖ
⫺3.25 0.009
201
202
203 10149⬘46.3ⴖ
⫺8.72ⴖ
⫺3.25 0.110
202
203
3 17649⬘01.0ⴖ
⫺8.98ⴖ
⫺3.25 0.109
203
3
2
859⬘51.1ⴖ
⫺4.91ⴖ
⫺3.25 0.054
2
1
3 31648⬘02.8ⴖ
2.29ⴖ
0.57 0.410
3
5
4 32417⬘43.8ⴖ
⫺0.19ⴖ
⫺0.19 0.016
6
5
3 33836⬘37.0ⴖ
⫺1.51ⴖ
⫺0.26 0.302
1
5
3 26849⬘43.7ⴖ
11.20ⴖ
1.57 0.528
2
5
3 31820⬘51.1ⴖ
⫺3.44ⴖ
⫺0.59 0.691
2
3
1
5107⬘14.4ⴖ
3.45ⴖ
0.68 0.497
2
3
5
9809⬘22.0ⴖ ⫺14.55ⴖ
⫺1.71 0.680
2
3
6
7142⬘48.5ⴖ
⫺2.97ⴖ
⫺0.31 0.392
2
3
4 16732⬘29.5ⴖ
1.48ⴖ
0.19 0.294
21.7
DATA SET WITH BLUNDERS
427
****************************************
Adjustment Statistics
****************************************
Iterations ⫽ 4
Redundancies ⫽ 12
Reference Variance ⫽ 1.316
Reference So ⫽ Ⳳ1.1
Possible blunder in observations with Std.Res. ⬎ 4
Convergence!
From analysis of the results, all observations containing blunders appear
to have been removed. However, it should also be noted that several remaining
distance and angle observations have very low redundancy numbers. This
identifies them as unchecked observations, which is also an undesirable situation. Thus, good judgment dictates reobservation of the measurements deleted. This weakness can also be seen in the size of the standard error ellipses
for the stations shown in Figure 21.7.
Note, especially, rotation of the error ellipses. That is, the uncertainty is
primarily in a direction perpendicular to the line to stations 1 and 102. This
condition is predictable since the angle 102–103–1 has been removed from
the data set. Furthermore, the crew on the northern leg never observed an
Figure 21.7 Standard error ellipse data for Example 21.1.
428
BLUNDER DETECTION IN HORIZONTAL NETWORKS
angle at station 1 that would tie into station 103, and thus the position of
station 103 was found by the intersection of two distances that nearly form
a straight line. This results in a larger error in the direction perpendicular to
the lines at this station.
This example demonstrates the process used to statistically detect and remove observational blunders. Whether the observations should be remeasured
depends on the intended use of the survey. Obviously, additional observations
would strengthen the network and probably reduce the size of the error
ellipses.
Observations between stations 102 and 201 also contribute to the overall
strength in the network. However, because a building obstructs that line, these
observations could not be obtained. This is a common problem in network
design. That is, it is sometimes physically impossible to gather observations
that would contribute to the total network strength. Thus, a compromise must
be made between the ideal network and what is physically obtainable. Balancing these aspects requires careful planning before the observations are
collected. Of course, line obstructions that occur due to terrain, vegetation,
or buildings, can now be overcome by using GPS.
21.8
SOME FURTHER CONSIDERATIONS
Equation (21.14) shows the relationship between blunders and their effects
on residuals as vi ⫽ ⫺ri li. From this relationship note that the effect of
the blunder, li, on the residual, vi, is directly proportional to the redundancy
number, ri. Therefore:
1. If ri is large (⬇ 1), the blunder greatly affects the residual and should
be easy to find.
2. If ri is small (⬇ 0), the blunder has little affect on the residual and will
be difficult to find.
3. If ri ⫽ 0, the blunder is undetectable and the parameters will be incorrect
since the error has not been detected.
Since redundancy numbers can range from 0 to 1, it is possible to compute
the minimum detectable error for a single blunder. For example, suppose that
a value of 4.0 is used to isolate observational blunders. Then, if the reference
variance of the adjustment is 6, all observations that have standardized residuals greater than 24.0 (4.0 ⫻ 6) are possible blunders. However, from Equation (21.14), it can be seen that for an observation with a redundancy number
of 0.2 (ri ⫽ 0.2) and a standardized residual of vi ⫽ 24.0, the minimum
detectable error is 24.0/0.2, or 120! Thus, a blunder, li, in this observation
as large as five times the desired level can go undetected due to its low
21.8
SOME FURTHER CONSIDERATIONS
429
redundancy number. This situation can be extended to observations that have
no observational checks; that is, ri is 0. In this case, Equation (21.14) shows
that it is impossible to detect any blunder, li, in the observation since vi /
ri is indeterminate.
With this taken into consideration, it has been shown that a marginally
detectable blunder in an individual observation is
li ⫽ S
冪qw
0
ii
2
ii
(21.19)
where 0 is the mean of the noncentral normal distribution shown in Figure
21.5, known as the noncentrality parameter. This parameter is the translation
of the normal distribution that leads to rejection of the null hypothesis, whose
values can be taken from nomograms developed by Baarda (1968). The sizes
of the values obtained from Equation (21.19) provide a clear insight into weak
areas of the network.
21.8.1
Internal Reliability
Internal reliability is found by examining how well observations check geometrically with each other. As mentioned previously, if a station is determined
uniquely, qii will be zero in Equation (21.19), and the computed value of li
is infinity. This indicates the lack of measurement self-checking. Since Equation (21.19) is independent of the actual observations, it provides a method
of detecting weak areas in networks. To minimize the sizes of the undetected
blunders in a network, the redundancy numbers of the individual observations
should approach their maximum value of 1. Furthermore, for uniform network
strength, the individual redundancy numbers, ri, should be close to the global
relative redundancy of r/m, where r is the number of redundant observations
and m is the number of observations in the network. Weak areas in the network are located by finding regions where the redundancy numbers become
small in comparison to relative redundancy.
21.8.2
External Reliability
An undetected blunder of size li has a direct effect on the adjusted parameters. External reliability is the effect of the undetected blunders on these
parameters. As li (a blunder) increases, so will its effect on X. The value
of X is given by
X ⫽ (ATWA)⫺1ATW ε
(21.20)
Again, this equation is datum independent. From Equation (21.20) it can
be seen that to minimize the value of Xi, the size of redundancy numbers
430
BLUNDER DETECTION IN HORIZONTAL NETWORKS
must be increased. Baarda suggested using average coordinate values in determining the effect of an undetected blunder with the following equation
⫽ X T Qxx X
(21.21)
where represents the noncentrality parameter.
The noncentrality parameter should remain as small as possible to minimize the effects of undetected blunders on the coordinates. Note that as the
redundancy numbers on the observations become small, the effects of undetected blunders become large. Thus, the effect of the coordinates of a station
from a undetected blunder is greater when the redundancy number is low. In
fact, an observation with a high redundancy number is likely to be detected
as a blunder.
A traverse sideshot can be used to explain this phenomenon. Since the
angle and distance to the station are unchecked in a sideshot, the coordinates
of the station will change according to the size of the observational blunders.
The observations creating the sideshot will have redundancy numbers of zero
since they are geometrically unchecked. This situation is neither good nor
acceptable in a well-designed observational system. In network design, one
should always check the redundancy numbers of the anticipated observations
and strive to achieve uniformly high values for all observations. Redundancy
numbers above 0.5 are generally sufficient to provide well-checked observations.
21.9
SURVEY DESIGN
In Chapters 8 and 19, the topic of observational system design was discussed.
Redundancy numbers can now be added to this discussion. A well-designed
network will provide sufficient geometric checks to allow observational blunders to be detected and removed. In Section 21.8.1 it was stated that if blunder
removal is to occur uniformly throughout the system, the redundancy numbers
should be close to the system’s global relative redundancy. Furthermore, in
Section 21.8.2 it was noted that redundancy numbers should be greater than
0.5. By combining these two additional concepts with the error ellipse sizes
and shapes, and stochastic model planning, an overall methodology for designing observational systems can be obtained.
To begin the design process, the approximate positions for stations to be
included in the survey must be determined. These locations can be determined
from topographic maps, photo measurements, or previous survey data. The
approximate locations of the control stations should be dictated by their desired locations, the surrounding terrain, vegetation, soils, sight-line obstructions, and so on. Field reconnaissance at this phase of the design process is
generally worthwhile to verify sight lines and accessibility of stations. Moving
21.9
SURVEY DESIGN
431
a station only a small distance from the original design location may greatly
enhance visibility to and from other stations but not change the geometry of
the network significantly. By using topographic maps in this process, sightline ground clearances can be checked by constructing profiles between
stations.
When approximate station coordinates are determined, a stochastic model
for the observational system can be designed following the procedures discussed in Chapter 7. In this design process, considerations should be given
to the abilities of the field personnel, quality of the equipment, and observational procedures. After the design is completed, specifications for field crews
can be written based on these design parameters. These specifications should
include the type of instrument used, number of turnings for angle observations, accuracy of instrument leveling and centering, misclosure requirements,
and many other items.
Once the stochastic model is designed, simulated observations are computed from the station coordinates, and a least squares adjustment of the
observations is to be done. Since actual observations have not been made,
their values are computed from the station coordinates. The adjustment will
converge in a single iteration, with all residuals equaling zero. Thus, the
reference variance must be assigned the a priori value of 1 to compute the
error ellipse axes and coordinate standard deviations. Having completed the
adjustment, the network can be checked for geometrically weak areas, unacceptable error ellipse sizes or shapes, and so on. This inspection may dictate
the need for any or all of the following: (1) more observations, (2) different
observational procedures, (3) different equipment, (4) more stations, (5) different network geometry, and so on. In any event, a clear picture of results
obtainable from the observational system will be provided by the simulated
adjustment and additional observations, or different network geometry can be
used.
It should be noted that what is expected from the design may not actually
occur, for numerous and varied reasons. Thus, systems are generally overdesigned. However, this tendency to overdesign should be tempered with the
knowledge that it will raise the costs of the survey. Thus, a balance should
be found between the design and costs. Experienced surveyors know what
level of accuracy is necessary for each job and design observational systems
to achieve the accuracy desired. It would be cost prohibitive and foolish always to design an observational system for maximum accuracy regardless of
the survey’s intended use. As an example, the final adjustment of the survey
in Section 21.7 had sufficient accuracy to be used in a mapping project with
a final scale of 1⬊1200 since the largest error ellipse semimajor axis (0.138
ft) would only plot as 0.0014 in. and is thus less than the width of a line on
the final map product.
For convenience, the steps involved in network design are summarized
below.
432
BLUNDER DETECTION IN HORIZONTAL NETWORKS
Step 1: Using a topographic map or aerial photos, lay out possible positions
for stations.
Step 2: Use the topographic map together with air photos to check sight lines
for possible obstructions and ground clearance.
Step 3: Do field reconnaissance, checking sight lines for obstructions not
shown on the map or photos, and adjust positions of stations as necessary.
Step 4: Determine approximate coordinates for the stations from the map or
photos.
Step 5: Compute values of observations using the coordinates from step 4.
Step 6: Using methods discussed in Chapter 6, compute the standard deviation of each observation based on available equipment and field measuring
procedures.
Step 7: Perform a least squares adjustment, to compute observational redundancy numbers, standard deviations of station coordinates, and error ellipses at a specified percent probability.
Step 8: Inspect the solution for weak areas based on redundancy numbers
and ellipse shapes. Add or remove observations as necessary, or reevaluate
measurement procedures and equipment.
Step 9: Evaluate the costs of the survey, and determine if some other method
of measurement (e.g., GPS) may be more cost-effective.
Step 10: Write specifications for field crews.
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
21.1
Discuss the effects of a distance blunder on a traverse closure and
explain how it can be identified.
21.2
Discuss the effects of an angle blunder on a traverse closure, and
explain how it can be identified.
21.3
Explain why a well-designed network has observational redundancy
numbers above 0.5 and approximately equal.
21.4
Create a list of items that should be included in field specifications
for a crew in a designed network.
21.5
Summarize the general procedures used in isolating observational
blunders.
433
PROBLEMS
21.6
How are control problems isolated in an adjustment?
21.7
Discuss possible causes for control problems in an adjustment.
21.8
Why is it recommended that there be at least three control stations in
a least squares adjustment?
21.9
Outline the procedures used in survey network design.
21.10 Using the procedures discussed in this chapter, analyze the data in
Problem
Figure P21.11
21.11 In Figure P21.11 the following data were gathered. Assuming that
the control stations have a published precision of 1⬊20,000, apply the
procedures discussed in this chapter to isolate and remove any apparent blunders in the data.
Approximate station
coordinates
Control stations
Station
Easting
Northing
Station
Easting
Northing
A
D
982.083
2686.270
1000.204
58.096
B
C
E
F
2507.7
4999.9
1597.6
2501.0
2500.6
998.6
200.0
1009.6
Distance observations
From
To
Distance (m)
S (m)
From
To
Distance (m)
S (m)
A
A
B
B
C
B
F
C
F
D
2139.769
1518.945
2909.771
1491.007
2497.459
0.023
0.021
0.025
0.021
0.023
E
E
D
D
C
F
A
F
E
F
1231.086
1009.552
969.386
1097.873
2498.887
0.021
0.020
0.020
0.021
0.023
434
BLUNDER DETECTION IN HORIZONTAL NETWORKS
Angle observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
F
B
F
C
F
D
F
E
A
F
A
B
C
D
E
A
A
B
B
C
C
D
D
E
E
F
F
F
F
F
E
F
A
F
B
F
C
F
F
D
B
C
D
E
A
5247⬘12.3ⴖ
4410⬘04.6ⴖ
4513⬘12.5ⴖ
5910⬘54.5ⴖ
2730⬘00.0ⴖ
2222⬘28.2ⴖ
7853⬘43.8ⴖ
7133⬘18.7ⴖ
8542⬘04.9ⴖ
4917⬘32.8ⴖ
9036⬘40.5ⴖ
8959⬘37.6ⴖ
7843⬘47.8ⴖ
5909⬘10.5ⴖ
4130⬘43.3ⴖ
3.4
2.8
2.8
2.7
2.4
2.5
3.3
3.8
3.7
3.4
3.1
2.8
3.3
3.6
3.1
21.12 Apply to procedures discussed in this chapter to isolate any blunders
in the following data.
Approximate coordinates
Control stations
Station
X (ft)
Y (ft)
Station
X (ft)
Y (ft)
A
H
44,680.85
169,721.77
78,314.23
214,157.12
B
C
D
E
F
G
21,112.93
49,263.50
107,855.97
113,747.95
186,743.04
206,710.31
151,309.33
213,175.12
145,744.68
78,968.90
78,314.24
138,216.04
Distance observations
Course
Distance (ft)
S (ft)
Course
Distance (ft)
S (ft)
AB
CD
EF
GH
AE
DH
DG
76,705.18
89,330.51
72,998.23
84,470.07
69,070.47
92,236.72
99,140.58
0.23
0.27
0.22
0.25
0.21
0.28
0.30
BC
DE
FG
CH
AD
BD
DF
67,969.04
67,035.17
63,141.95
120,461.89
92,401.26
86,921.33
103,778.98
0.20
0.20
0.19
0.36
0.28
0.26
0.31
PROBLEMS
435
Angle Observations
Backsight
Occupied
Foresight
Angle
S (ⴖ)
B
D
C
B
H
D
G
D
F
D
E
A
D
D
A
B
C
H
G
F
E
A
B
B
C
C
H
H
G
G
F
F
E
E
A
D
D
D
D
D
D
D
D
A
D
D
D
C
D
H
D
G
D
D
F
E
B
C
H
G
F
E
A
6101⬘40ⴖ
6826⬘08ⴖ
6912⬘11ⴖ
6527⬘20ⴖ
4928⬘44ⴖ
4724⬘34ⴖ
6805⬘33ⴖ
5940⬘32ⴖ
7555⬘10ⴖ
6754⬘48ⴖ
4000⬘31ⴖ
8530⬘03ⴖ
9533⬘32ⴖ
4619⬘22ⴖ
5032⬘12ⴖ
4520⬘27ⴖ
8306⬘44ⴖ
5213⬘51ⴖ
3610⬘04ⴖ
4426⬘04ⴖ
4810⬘35ⴖ
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.3
2.1
2.1
2.1
2.1
2.1
2.1
2.1
2.1
21.13 Using the data set Blunder.dat on the CD that accompanies this book,
isolate any blunders that are detectable with a rejection criteria of
3.29.
Figure P21.14
21.14 As shown in Figure P21.14, the following approximate station coordinates were determined from a map of an area where a second-order
class I survey (1⬊50,000) is to be performed. All sight lines to neighboring stations have been checked and are available for conventional
436
BLUNDER DETECTION IN HORIZONTAL NETWORKS
observations of distances, angles, and azimuths. The control stations
are visible to their nearest neighbors. Design a control network that
will meet the specified accuracy at a 95% confidence level and have
sufficient checks to ensure the reliability of the results. Stations J218,
J219, and ROCK are first-order control stations (1⬊100,000).
Approximate station
locations
Control stations
Station
X (m)
Y (m)
Station
X (m)
Y(m)
J218
J219
ROCK
283,224.223
154,995.165
521,287.251
116,202.946
330,773.314
330,276.310
101
102
103
104
105
106
107
108
280,278
276350
360,147
356,219
352,291
455,728
446,563
440,671
194,109
278,887
121,768
195,090
274,304
132,570
198,036
270,700
21.15 Using the criteria in Problem 21.14, design a carrier-phase GPS network to establish the control network. Assume that all stations have
no obstructions above 15.
Practical Problems
21.16 Design a 6 mi ⫻ 6 mi control network having a minimum of eight
control stations using a topographic map of your local area. Design
a traditional measurement network made up of angles, azimuths, and
distances so that the largest ellipse axis at a 95% confidence level is
less than 0.20 ft and so that all observations have redundancy numbers
greater than 0.5. In the design, specify the shortest permissible sight
distance, the largest permissible errors in pointing, reading, and instrument and target setup errors, the number of repetitions necessary
for each angle measurement, and the necessary quality of angle and
distance measuring instruments. Use realistic values for the instruments. Plot profiles of sight lines for each observation.
21.17 Design a 6 mile ⫻ 6 mile GPS control network to be established by
differential GPS which has a minimum of eight control stations, using
a topographic map of your local area to select station locations. Design the survey so that all baseline observations included in the network have redundancy numbers greater than 0.5. In the design, use a
unit matrix for the covariance matrix of the baselines.
CHAPTER 22
GENERAL LEAST SQUARES METHOD
AND ITS APPLICATION TO CURVE
FITTING AND COORDINATE
TRANSFORMATIONS
22.1
INTRODUCTION TO GENERAL LEAST SQUARES
When fitting points to a straight line, it must be recognized that both the x
and y coordinates contain errors. Yet in the mathematical model presented in
Section 11.11.1, illustrated in Figure 11.2, the residuals are applied only to
the y coordinate. Because both coordinates contain errors, this mathematical
model fails to account for the x coordinate being a measurement. In this
chapter the general least squares method is presented and its use in performing adjustments where the observation equations involve more than a single
measurement is demonstrated.
22.2 GENERAL LEAST SQUARES EQUATIONS FOR FITTING A
STRAIGHT LINE
Consider the data illustrated in Figure 11.2. To account properly for both the
x and y coordinates being measurements, the observation equation must contain residuals for both measurements. That is, Equation (11.40) must be rewritten as
F(x,y) ⫽ (y ⫹ vy) ⫺ m(x ⫹ vx) ⫺ b ⫽ 0
(22.1)
In Equation (22.1), x and y are a point’s coordinate pair with residuals vx and
vy, respectively, m is the slope of the line, and b is the y intercept. Equation
(22.1) contains vx, vy, m, and b as unknowns and is nonlinear. Thus, its soAdjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
437
438
GENERAL LEAST SQUARES METHOD
lution is obtained by using the methods outlined in Section 11.10. The resulting linearized form of Equation (22.1) is
F
F
F
F
v ⫹
v ⫹
dm ⫹
db ⫽ ⫺(m0 x ⫹ b0 ⫺ y)
x x
y y
m
b
(22.2)
where the partial derivatives are
F
⫽ ⫺m
x
F
⫽1
y
F
⫽ ⫺x
m
F
⫽ ⫺1
b
(22.3)
For the four data points A, B, C, and D, substituting into Equations (22.3)
into (22.2) the following four equations can be written:
⫺m0vxA ⫹ vyA ⫺ xAdm ⫺ db ⫽ ⫺(m0 xA ⫹ b0 ⫺ yA)
⫺m0vxB ⫹ vy ⫺ xBdm ⫺ db ⫽ ⫺(m0 xB ⫹ b0 ⫺yB)
(22.4)
⫺m0vxC ⫹ vyC ⫺ xCdm ⫺db ⫽ ⫺(m0 xC ⫹ b0 ⫺ yC)
⫺m0vxD ⫹ vyD ⫺ xDdm ⫺ db ⫽ ⫺(m0 xD ⫹ b0 ⫺ yD)
In matrix form, Equations (22.4) can be written as
(22.5)
BV ⫹ JX ⫽ K
where
B⫽
V⫽
冤冥
⫺m0
0
0
0
vxA
vyA
vxB
vyB
vxC
vyC
vxD
vyD
1
0
0 ⫺m0
0
0
0
0
0
0
1
0
0 ⫺m0
0
0
冋 册
dm
X⫽
db
0
0
0
0
1
0
0 ⫺m0
K⫽
冤
0
0
0
0
⫺xA
⫺xB
⫺xC
⫺xD
J⫽
⫺1
⫺1
⫺1
⫺1
冥
⫺(m0 xA ⫹ b0 ⫺ yA)
⫺(m0 xB ⫹ b0 ⫺ yB)
⫺(m0 xC ⫹ b0 ⫺ yC)
⫺(m0 xD ⫹ b0 ⫺ yD)
(22.6)
Now since both x and y are measured coordinates, they may each have
individual estimated standard errors. Assuming that the coordinates are from
independent observations, the four points will have eight measured coordinates and a cofactor matrix of
22.3
2xA
Q⫽
22.3
0
0
0
0
0
0
0
0
2yA
0
0
0
0
0
0
0
0
2xB
0
0
0
0
0
0
0
0
2yB
0
0
0
0
GENERAL LEAST SQUARES SOLUTION
0
0
0
0
2xC
0
0
0
0
0
0
0
0
2yC
0
0
0
0
0
0
0
0
2xD
0
0
0
0
0
0
0
0
439
2yD
GENERAL LEAST SQUARES SOLUTION
In solving the general least squares problem, an equivalent solution is
achieved. For this solution, the following equivalent weight matrix is created
for the system:
We ⫽ (BQBT)⫺1
(22.7)
where B is as defined in Equation (22.6). Using the equivalent weight matrix
in Equation (22.7), the equivalent matrix system is
J TWe JX ⫽ J TWeK
(22.8)
Equation (22.8) has the solution
X ⫽ (J TWe J)⫺1J TWeK
(22.9)
Since this is a nonlinear equation system, the corrections in matrix X are
applied to the initial approximations, and the method is repeated until the
system converges. The equivalent residuals vector Ve is found following the
usual procedure of
Ve ⫽ JX ⫺ K
(22.10)
Using Equation (22.10), the observational residuals are
V ⫽ QBTWeVe
(22.11)
Also, since this is a nonlinear problem and the observations are being
adjusted, the observations should also be updated according to their residuals.
Thus, the updated observations for the second iteration are
l⬘i ⫽ li ⫹ vi
(22.12)
where l⬘i are the second iterations observations, li the original observations,
and vi the observations’ corresponding residuals. Generally in practice, the
440
GENERAL LEAST SQUARES METHOD
original observations are ‘‘close’’ to their final adjusted values, and thus Equation (22.12) is not actually used since the second iteration is only a check for
convergence.
Finally, the reference variance for the adjustment can be computed using
the equivalent residuals and weight matrix employing the equation
S02 ⫽
V Te Weve
r
(22.13)
where r is the number of redundancies in the system.
It should be noted that the same results can be obtained using the observational residuals and the following expression:
S20 ⫽
V TWV
r
(22.14)
Example 22.1 Numerical Solution of the Straight-Line-Fit Problem.
Recall the least squares fit of points to a line in Section 11.11.1. In that
example, the measured coordinate pairs were
A: (3.00, 4.50)
B: (4.25, 4.35)
C: (5.50, 5.50)
D: (8.00, 5.50)
and the solution for the slope of the line, m, and y intercept, b, were
m ⫽ 0.246
(a)
b ⫽ 3.663
Additionally, the residuals were
⫺0.10
V ⫽ AX ⫺ L ⫽
0.46
⫺0.48
0.13
Now this problem will be solved using the general least squares method.
Assume that the following Q matrix is given:
0.0202
0
0
0.0152
0
0
0
0
Q⫽
0
0
0
0
0
0
0
0
0
0
0.0232
0
0
0
0
0
0
0
0
0.0362
0
0
0
0
0
0
0
0
0.0332
0
0
0
0
0
0
0
0
0.0282
0
0
0
0
0
0
0
0
0.0162
0
0
0
0
0
0
0
0
0.0192
22.3
441
GENERAL LEAST SQUARES SOLUTION
SOLUTION The step-by-step procedure for solving this problem using general least squares follows.
Step 1: Compute the initial approximations. Initial approximations for both
m and b are found by using two points and solving the unique system. For
this example the values from the solution of Section 11.11.1 given above
will be used.
Step 2: Develop the appropriate matrices. In accordance with Equation (22.6),
the B matrix is
⫺0.246
B⫽
0
0
0
1
0
0
0
0
0
0 ⫺0.246 1
0
0
0
0
0
0 ⫺0.246 1
0
0
0
0
0
0 ⫺0.246
0
0
0
1
Using Equation (22.7), the equivalent weight matrix is
We ⫽ (BQBT)⫺1
1
⫽
10,000
2.5
0
0
0
0 13.3 0
0
0
0
8.5 0
0
0
0 3.8
4012.7
0
0
0
0
753.0
0
0
⫽
0
0
1176.6
0
0
0
0
2656.1
Step 3: Solve the system. The first iteration corrections are found using Equation (22.9) as
X ⫽ (J TWe J)⫺1J TWe K ⫽
冋
册
⫺0.0318
0.1907
where the J, K, and J TWe J matrices for the first iteration were
J⫽
冤
⫺3.00
⫺4.25
⫺5.50
⫺8.00
J TWe J ⫽
冥
⫺1
⫺1
⫺1
⫺1
冋
冤
0.246(3.00)
0.246(4.25)
K⫽
0.246(5.50)
0.246(8.00)
册
255,297.91 42,958.45
42,958.45 8,598.40
⫹
⫹
⫹
⫹
3.663
3.663
3.663
3.663
⫺
⫺
⫺
⫺
冥冤 冥
4.50
⫺0.099
4.25
0.458
⫽
5.50
⫺0.484
5.50
0.131
J TWe K ⫽
冋
册
72.9737
273.5321
Step 4: Apply the corrections to m0 and b0 to update the initial approximations for the second iteration,
442
GENERAL LEAST SQUARES METHOD
m ⫽ 0.246 ⫺ 0.0318 ⫽ 0.2142
b ⫽ 3.663 ⫹ 0.1907 ⫽ 3.8537
Second iteration. During the second iteration, only the unknown parameters
are updated, and thus only B, We, and K matrices differ from their first iteration counterparts. Their second iteration values are
⫺0.2142
B⫽
0
0
0
1
0
0
0
0
0
0
0 ⫺0.2142 1
0
0
0
0
0
0
0 ⫺0.2142 1
0
0
0
0
0
0
0 ⫺0.2142 1
4109.3
0
0
0
0
757.4
0
0
We ⫽
0
0
199.1
0
0
0
0
2682.8
⫺0.00370
K⫽
0.51405
⫺0.46820
0.06730
The corrections after this iteration are
X⫽
冋
册
0.00002
0.00068
Thus, m and b are 0.214 and 3.854 to three decimal places. Using Equation
(22.10), the equivalent residual vector is
冤 冥
⫺0.0030
Ve ⫽
0.5148
⫺0.4674
0.0681
Using Equation (22.11), the observation residual vector is
Figure 22.1 General least squares fit of points to a line.
22.4
TWO-DIMENSIONAL COORDINATE TRANSFORMATION BY GENERAL LEAST SQUARES
443
Figure 22.2 Residuals for point C.
0.0010
⫺0.0027
⫺0.0442
V ⫽ QBTWeVe ⫽
0.5053
0.1307
⫺0.4394
⫺0.0100
0.0660
A graphical interpretation of the residuals is shown in Figure 22.1. Notice
how the equivalent residuals are aligned with the y axis, and that the observational residuals exist in the primary x- and y-axis directions. These residuals
are shown more clearly in Figure 22.2, which is an enlarged view of the
portion of Figure 22.1 that surrounds point C. The equivalent residual of C
is ⫺0.4674 from the line, and the observational residuals vxc ⫽ 0.1307 and
vyc ⫽ ⫺0.4394 are parallel to the x and y axes, respectively. This general
solution is more appropriate for adjusting the actual coordinate measurements.
Note that this solution is slightly different from that determined in Example
11.3.
Using Equation (22.13), the reference standard deviation is
S0 ⫽
⫽ 15.4
冪4475.2
⫺2
22.4 TWO-DIMENSIONAL COORDINATE TRANSFORMATION BY
GENERAL LEAST SQUARES
As presented in Chapter 18, two-dimensional coordinate transformations are
commonly used to convert points from one two-dimensional coordinate system to another. Again, the general least squares method is a more appropriate
method for these transformations since the coordinates in both systems are
measurements that contain errors.
444
GENERAL LEAST SQUARES METHOD
22.4.1
Two-Dimensional Conformal Coordinate Transformation
The two-dimensional conformal coordinate transformation, presented in
Chapter 18, has four unknowns, consisting of a scale factor, rotation angle,
and two translations. Equations (18.5) express this transformation, and they
are repeated here for convenience. The transformation equations are
X ⫽ ax ⫺ by ⫹ c
(22.15)
Y ⫽ bx ⫹ ay ⫹ d
Equations (22.15) can be rearranged as
F: ax ⫺ by ⫹ c ⫺ X ⫽ 0
(22.16)
G: bx ⫹ ay ⫹ d ⫺ Y ⫽ 0
Since the coordinates from both the xy and XY systems contain errors, Equations (22.16) are rewritten as
F(x, y, X, Y) ⫽ a(x ⫹ vx) ⫺ b(y ⫹ vy) ⫹ c ⫺ (X ⫹ vx) ⫽ 0
G(x, y, X, Y) ⫽ b(x ⫹ vx) ⫹ a(y ⫹ vy) ⫹ d ⫺ (Y ⫹ vY) ⫽ 0
(22.17)
These equations are nonlinear in terms of their observations and residuals.
They are solved by linearizing the equations and iterating to a solution. The
partial derivatives with respect to each unknown are
ƒ
⫽a
x
ƒ
⫽x
a
ƒ
ƒ
g
g
g
⫽ ⫺b
⫽ ⫺1
⫽b
⫽a
⫽ ⫺1
y
X
x
y
Y
ƒ
ƒ
g
g
g
⫽ ⫺y
⫽1
⫽y
⫽x
⫽1
b
c
a
b
d
(22.18)
Using the partial derivatives in Equation (22.18), a matrix for each point can
be built as
冋
a0
b0
⫽
⫺b0
a0
冋
1
0
0
1
册冤
冥
vx
vy
vX
vY
冋
册
册冤
x ⫺y 1 0
⫹
y
x 0 1
X ⫺ (a0 x ⫹ b0 y ⫹ Tx)
Y ⫺ (b0 x ⫹ a0 y ⫹ Ty)
冥
da
db
dc
dd
(22.19)
For a redundant system, the matrices are solved following the matrix procedures outlined in Section 22.3.
22.4
TWO-DIMENSIONAL COORDINATE TRANSFORMATION BY GENERAL LEAST SQUARES
445
Example 22.2 Four fiducial points are digitized from an aerial photo, and
their measured (x,y) and control (X,Y) values are listed in Table 22.1. The
standard deviations of these measurements are also listed. What are the most
probable values for the transformation parameters and the resulting residuals?
SOLUTION In this problem, initial approximations for a, b, c, and d must
first be computed. These values can be found using a standard least squares
adjustment, or by solving the system with only two points. The former procedure was demonstrated in Example 18.1. Using standard least squares, initial approximations for the parameters are determined to be
a0 ⫽ 25.386458
b0 ⫽ ⫺0.8158708
c0 ⫽ ⫺137.216
d0 ⫽ ⫺150.600
Now the B, Q, J, and K matrices described in Section 21.3 can be formed.
They are listed below (note that the numbers are rounded to three decimal
places for publication purposes only):
25.386
0.816 ⫺1 0
0
0 0 0
0
0 0 0
0
0 0 0
0 ⫺1
0
0 0 0
0
0 0 0
0
0 0 0
0
0 0 0 25.386 0.816 ⫺1 0
0
0 0 0
0
0 0 0
0
0 0 0 ⫺0.816 25.386 0 ⫺1
0
0 0 0
0
0 0 0
0
0 0 0
0
0 0 0 25.386 0.816 ⫺1 0
0
0 0 0
0
0 0 0
0
0 0 0 ⫺0.816 25.386 0 ⫺1
0
0 0 0
0
0 0 0
0
0 0 0
0
0 0 0 25.386 0.816 ⫺1 0
0
0 0 0
0
0 0 0
0
0 0 0 ⫺0.816 25.386 0 ⫺1
⫺0.816 25.386
B⫽
0.764 ⫺5.960 1 0
5.960
0.764 0 1
5.062 ⫺10.541 1 0
10.541
5.062 0 1
J⫽
9.663 ⫺6.243 1 0
6.243
9.663 0 1
5.350 ⫺1.654 1 0
1.654
5.350 0 1
⫺0.03447270
⫺0.08482509
0.11090026
0.13190015
K⫽
⫺0.18120912
⫺0.00114251
0.04999940
⫺0.02329275
TABLE 22.1 Data for a Two-Dimensional Conformal Coordinate
Transformation
Point
1
3
5
7
X SX
⫺113.000
0.001
112.998
0.001
Y SY
0.002
0.003
0.002
112.993
0.002
0.003
0.002 ⫺112.999
x sx
0.002
0.002
0.002
0.002
0.7637
5.0620
9.6627
5.3500
y sy
0.026 5.9603 0.028
0.024 10.5407 0.030
0.028 6.2430 0.022
0.024 1.6540 0.026
446
GENERAL LEAST SQUARES METHOD
Also, the Q matrix is
0.0262
0.0282
0.0022
0.0022
0.0242
0.0302
0.0022
0.0022
Q⫽
0.0282
0.0222
0.0022
0.0022
0.0242
0.0262
0.0022
0.0022
The solution for the first iteration is
X⫽
⫺0.000124503
⫺0.000026212
⫺0.000325016
⫺0.000029546
Adding these corrections to the initial approximations, yields
a ⫽ 25.38633347
b ⫽ ⫺0.815897012
c ⫽ ⫺137.2163
d ⫽ ⫺150.6000
In the next iteration, only minor corrections occur, and thus the system has
converged to a solution. The residuals and reference variance are computed
as before. Although the solution has changed only slightly from the standard
least squares method, it properly considers the fact that each observation
equation contains four measurements. Of course, once the transformation parameters have been determined, any points that exist only in the xy system
can be transformed into the XY system by substitution into Equations (22.15).
This part of the problem is not demonstrated in this example. For the remainder of the chapter, only the adjustment model is developed.
22.4
TWO-DIMENSIONAL COORDINATE TRANSFORMATION BY GENERAL LEAST SQUARES
22.4.2
447
Two-Dimensional Affine Coordinate Transformation
As discussed in Section 17.5, the main difference between conformal and
affine transformations is that the latter allows for different scales along the x
and y axes and accounts for nonorthogonality in the axes. This results in six
parameters. Equations (18.9) express the affine transformation, and they are
repeated here for convenience.
X ⫽ ax ⫹ by ⫹ c
(22.20)
Y ⫽ dx ⫹ ey ⫹ ƒ
Equation (22.20) can be rewritten as
F(x,y,X,Y) ⫽ ax ⫹ by ⫹ c ⫺ X ⫽ 0
(22.21)
G(x,y,X,Y) ⫽ dx ⫹ ey ⫹ ƒ ⫺ Y ⫽ 0
Again, Equation (22.21) consists of observations in both the x and y coordinates, and thus it is more appropriate to use the general least squares
method. Therefore, these equations can be rewritten as
F(x,y,X,Y) ⫽ a(x ⫹ vx) ⫹ b(y ⫹ vy) ⫹ c ⫺ (X ⫹ vX) ⫽ 0
G(x,y,X,Y) ⫽ d(x ⫹ vx) ⫹ e(y ⫹ vy) ⫹ ƒ ⫺ (Y ⫹ vY) ⫽ 0
(22.22)
For each point, the linearized equations in matrix form are
冋
a0 b0 ⫺1 0
d0 e0 0 ⫺1
⫽
冋
册冤
冥
vx
vy
vX
vY
⫹
冋
x y
0 0
册
X ⫺ (a0 x ⫹ b0 y ⫹ c0)
Y ⫺ (d0 x ⫹ e0 y ⫹ ƒ0)
1
0
0
x
冤冥
册
0 0
y 1
da
db
dc
dd
de
df
(22.23)
Two observation equations, like those of Equation (22.23), result for each
control point. Since there are six unknown parameters, three control points
are needed for a unique solution. With more than three, a redundant system
exists, and the solution is obtained following the least squares procedures
outlined in Section 22.3.
448
GENERAL LEAST SQUARES METHOD
22.4.3
Two-Dimensional Projective Transformation
The two-dimensional projective coordinate transformation converts a projection of one plane coordinate system into another nonparallel plane system.
This transformation was developed in Section 18.6. Equations (18.12) for this
transformation are repeated here for convenience.
X⫽
a1x ⫹ b1y ⫹ c1
a3x ⫹ b3y ⫹ 1
(22.24)
a x ⫹ b2 y ⫹ c2
Y⫽ 2
a3x ⫹ b3y ⫹ 1
Unlike the conformal and affine types, this transformation is nonlinear in
its standard form. In their general form, the projective equations become
F(x,y,X,Y) ⫽
a1(x ⫹ vx) ⫹ b1(y ⫹ vy) ⫹ c1
⫺ (X ⫹ vX) ⫽ 0
a3(x ⫹ vx) ⫹ b3(y ⫹ vy) ⫹ 1
a2(x ⫹ vx) ⫹ b2(y ⫹ vy) ⫹ c2
G(x,y,X,Y) ⫽
⫺ (Y ⫹ vY) ⫽ 0
a3(x ⫹ vx) ⫹ b3(y ⫹ vy) ⫹ 1
(22.25)
Again a linearized form for Equations (22.25) is needed. The partial derivatives for the unknown parameters were given in Section 18.6, and the
remaining partial derivatives are given below.
F
a (b y ⫹ 1) ⫺ a3(b1y ⫹ c1)
⫽ 1 3
x
(a3x ⫹ b3y ⫹ 1)2
F
b (a x ⫹ 1) ⫺ b3(a1y ⫹ c1)
⫽ 1 3
y
(a3x ⫹ b3y ⫹ 1)2
G
a (b y ⫹ 1) ⫺ a3(b2 y ⫹ c2)
⫽ 2 3
x
(a3 ⫹ b3y ⫹ 1)2
G
b (a x ⫹ 1) ⫺ b3(a2 y ⫹ c2)
⫽ 2 3
y
(a3x ⫹ b3y ⫹ 1)2
In matrix form, the linearized equations for each point are
F
x
G
x
F
⫺1 0
y
G
0 ⫺1
y
vx
vy
vX
vY
⫹
F F F
a1 b1 c1
0
0
F
a3
G G G G
0
a2 b2 c2 a3
0
0
0
F
b3
G
b3
da1
db1
dc1
da2
⫽K
db2
dc2
da3
db3
(22.26)
Equation (22.26) gives the observation equations for the two-dimensional
projective transformation for one control point and the K matrix is defined in
22.5
THREE-DIMENSIONAL CONFORMAL COORDINATE TRANSFORMATION
449
Equation (18.13). Since there are eight unknown parameters, four control
points are needed for a unique solution. More than four control points yields
a redundant system that can be solved following the steps outlined in Section
22.3.
22.5 THREE-DIMENSIONAL CONFORMAL COORDINATE
TRANSFORMATION BY GENERAL LEAST SQUARES
As explained in Section 18.7, this coordinate transformation converts points
from one three-dimensional coordinate system to another. Equations (18.15)
express this transformation, and the matrix form of those equations is
(22.27)
X ⫽ SRx ⫹ T
where the individual matrices are as defined in Section 18.7.
Equation (18.15) gives detailed expressions for the three-dimensional coordinate transformation. Note that these equations involve six observations,
xyz and XYZ. In general least squares, these equations can be rewritten as
ƒ(x,y,z,X,Y,Z) ⫽ S[r11(x ⫹ Vx) ⫹ r21(y ⫹ Vy) ⫹ r31(z ⫹ Vz)] ⫺ (X ⫹ VX) ⫽ 0
g(x,y,z,X,Y,Z) ⫽ S[r12(x ⫹ Vx) ⫹ r22(y ⫹ Vy) ⫹ r32(z ⫹ Vz)] ⫺ (Y ⫹ VY) ⫽ 0
h(x,y,z,X,Y,Z) ⫽ S[r13(x ⫹ Vx) ⫹ r23(y ⫹ Vy) ⫹ r33(z ⫹ Vz)] ⫺ (Z ⫹ VZ) ⫽ 0
(22.28)
Equations (22.28) are for a single point, and again, they are nonlinear.
They can be expressed in linearized matrix form as
ƒ
x
ƒ
y
ƒ
z
g
x
g
y
h
x
h
y
⫹
⫺1
0
0
g
z
0 ⫺1
0
h
z
0
0
⫺1
ƒ
S
0
ƒ
⭸2
ƒ
⭸3
1
⭸g
⭸S
⭸g
⭸1
⭸g
⭸2
⭸g
⭸3
0
⭸h
⭸S
⭸h
⭸1
⭸h
⭸2
⭸h
⭸3
0
vx
vy
vz
vX
vY
vZ
0
0
1
0
0
1
dS
d1
d2
0
d3 ⫽ 0
dTX
0
dTY
dTZ
(22.29)
450
GENERAL LEAST SQUARES METHOD
where
ƒ
⫽ Sr11
x
ƒ
⫽ Sr21
y
ƒ
⫽ Sr31
z
g
⫽ Sr12
x
g
⫽ Sr22
y
g
⫽ Sr32
z
h
⫽ Sr13
x
h
⫽ Sr23
y
h
⫽ Sr33
z
The remaining partial derivatives were given in Section 18.7. This system is
solved using the methods discussed in Section 22.3.
Example 22.3 Estimated errors were added to the control coordinates in
Example 18.4. The control data are repeated in Table 22.2 and standard deviations of the control coordinates, needed to form the Q matrix, are also
listed. Following Table 22.2, output from the program ADJUST is listed. Note
that the solution differs somewhat from the one obtained by standard least
squares in Example 18.4.
SOLUTION
3D Coordinate Transformation of File:
using generalized least squares method
========================================================
Measured Points
========================================================
NAME
x
y
z
Vx
Vy
Vz
========================================================
1 1094.883
820.085 109.821 ⫺0.001 ⫺0.001 ⫺0.000
2
503.891 1598.698 117.685 ⫺0.002
0.000
0.000
3 2349.343
207.658 151.387 ⫺0.001
0.000
0.000
4 1395.320 1348.853 215.261
0.001
0.000 ⫺0.001
TABLE 22.2 Control Data for the Three-Dimensional Conformal Coordinate
Transformation
Point
X
Y
Z
SX
SY
SZ
1
2
3
4
10,037.81
10,956.68
8,780.08
10,185.80
5262.09
5128.17
4840.29
4700.21
772.04
783.00
782.62
851.32
0.05
0.04
0.02
0.03
0.06
0.06
0.04
0.05
0.05
0.09
0.02
0.03
PROBLEMS
451
CONTROL POINTS
========================================================
NAME
X
VX
Y
VY
Z
VZ
========================================================
1 10037.810 ⫺0.063 5262.090 ⫺0.026 772.040 ⫺0.001
2 10956.680 ⫺0.019 5128.170
0.063 783.000 ⫺0.027
3 8780.080
0.000 4840.290
0.038 782.620 ⫺0.001
4 10185.800
0.032 4700.210 ⫺0.085 851.320
0.007
Transformation Coefficients
========================================================
Scale ⫽ 0.94996 ⫹/⫺ 0.00002
Omega ⫽ 2⬚ 17⬘ 00.0ⴖ ⫹ /⫺ 0⬚ 00⬘ 26.7ⴖ
Phi ⫽ ⫺0⬚ 33⬘ 05.6ⴖ ⫹ /⫺ 0⬚ 00⬘ 06.1ⴖ
Kappa ⫽ 224⬚ 32⬘ 11.5ⴖ ⫹ /⫺ 0⬚ 00⬘ 07.7ⴖ
Tx ⫽ 10233.855 ⫹/⫺ 0.066
Ty ⫽ 6549.964 ⫹/⫺ 0.055
Tz ⫽ 720.867 ⫹/⫺ 0.219
Reference Standard Deviation: 1.293
Degrees of Freedom: 5
Iterations: 3
Transformed Coordinates
========================================================
NAME
X
Sx
Y
Sy
Z
Sz
========================================================
1 10037.874 0.082 5262.116 0.063 772.041 0.275
2 10956.701 0.087 5128.106 0.070 783.027 0.286
3
8780.080 0.098 4840.251 0.087 782.622 0.314
4 10185.767 0.096 4700.296 0.072 851.313 0.324
5 10722.016 0.075 5691.210 0.062 766.067 0.246
6 10043.245 0.074 5675.887 0.060 816.857 0.246
PROBLEMS
Note: For problems requiring least squares adjustment, if a computer program
is not distinctly specified for use in the problem, it is expected that the least
squares algorithm will be solved using the program MATRIX, which is included on the CD supplied with the book.
22.1
Solve Problem 11.9 using the general least squares method. Assign
all coordinates a standard deviation of 1.
452
GENERAL LEAST SQUARES METHOD
22.2
Do Problem 11.11 using the general least squares method. Assign the
computed coordinates the following standard deviations.
Station
SX
SY
Station
SX
SY
A
C
E
G
I
0.001
0.020
0.020
0.020
0.020
0.001
0.013
0.027
0.035
0.046
B
D
F
H
0.020
0.020
0.020
0.020
0.012
0.021
0.028
0.040
22.3
Solve Problem 11.12 using the general least squares method. Assign
all coordinates a standard deviation of 0.1.
22.4
Do Example 18.1 using the general least squares method. Assume
estimated standard deviations for the coordinates of SX ⫽ SY ⫽ 0.05
and Sx ⫽ Sy ⫽ 0.005.
22.5
Do Problem 18.1 using the general least squares method. Assume
estimated standard deviations for all the coordinates of 0.05. Compare this solution with that of Problem 22.4.
22.6
Solve Problem 18.2 using the general least squares method. Assume
estimated standard deviations of 0.003 mm for all the control
coordinates.
22.7
Use the general least squares method to transform the points from the
measured system to the control system using a two-dimensional:
(a) conformal transformation.
(b) affine transformation.
(c) projective transformation.
Measured coordinates
Point x (mm)
6
7
3
4
5
32
22
12
13
23
33
y (mm)
103.55 ⫺103.670
0.390 ⫺112.660
0.275 111.780
103.450 102.815
112.490
⫺0.395
18.565 ⫺87.580
⫺5.790
2.035
6.840
95.540
86.840 102.195
93.770
2.360
92.655 ⫺90.765
Control coordinates
Sx (mm) Sy (mm) X (mm)
0.003
0.005
0.004
0.003
0.003
0.004
0.005
0.007
0.004
0.003
Y (mm)
103.951 ⫺103.969
0.001 ⫺112.999
0.001 112.993
103.956 103.960
112.998
0.003
SX (mm) SY (mm)
0.009
0.009
0.009
0.009
0.009
0.009
0.009
0.009
0.009
0.009
453
PROBLEMS
22.8
Given the following data, transform the points from the measured
system to the control system using a three-dimensional conformal
coordinate transformation. Assume that all coordinates have estimated
standard deviations of 0.05 m. Use the general least squares
method.
Measured coordinates
22.9
Control coordinates
Point
x (m)
y (m)
z (m)
X (m)
Y (m)
Z (m)
1
2
3
4
5
6
607.54
598.98
643.65
628.58
666.27
632.59
501.63
632.36
421.28
440.51
298.16
710.62
469.09
82.81
83.50
82.27
98.29
103.01
390.13
371.46
425.43
499.74
630.95
419.18
469.32
81.14
82.38
Solve Problem 18.14 using general least squares. Assume that all the
control coordinates have estimated standard deviations of 0.005 m.
22.10 Do Problem 22.1 using the ADJUST program.
Solve each problem with the ADJUST program using the standard and general
least squares adjustment options. Compare and explain any differences noted
in the solutions.
22.11 Problem 22.6
22.12 Problem 22.7
22.13 Problem 22.8
Programming Problems
Develop a computational program that constructs the B, Q, J, and K matrices
for each of the following transformations.
22.14 A two-dimensional conformal coordinate transformation.
22.15 A two-dimensional affine coordinate transformation.
22.16 A two-dimensional projective coordinate transformation.
22.17 A three-dimensional conformal coordinate transformation.
CHAPTER 23
THREE-DIMENSIONAL GEODETIC
NETWORK ADJUSTMENT
23.1
INTRODUCTION
With the advent of total station instruments, survey data are being collected
in three dimensions. Thus, it is advantageous to develop an adjustment model
that works in three dimensions. Rigorous triangulation adjustment models date
back to Bruns (1878). The main observational data consist of horizontal angles, vertical angles, azimuths, and slant distances. It is also possible to include differential leveling in the model. Since all data are collected on Earth’s
surface, the local geodetic coordinate system provides a natural system in
which to perform the adjustment.
As shown in Figure 23.1, the local geodetic system is oriented such that
the n axis points along the meridian of the ellipse (local geodetic north), the
u axis is aligned along the normal of the ellipsoid, and the e axis creates a
left-handed coordinate system. The local geodetic coordinate system can be
related to the geocentric coordinate system (see Section 17.5) through a series
of three-dimensional rotations discussed in Section 18.7. To align the X axis
with the e axis, the Z axis is rotated by an amount of ⫺ 180⬚. Then the Z
axis is aligned with the u axis by a rotation of ⫺ 90⬚ about the once-rotated
X axis. Changing the sign of e to convert to a right-handed coordinate system,
the resulting expression is
⌬n
⫺⌬e
⌬u
454
⌬X
⫽ R2 ( ⫺ 90⬚)R3( ⫺ 180⬚) ⌬Y
⌬Z
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
(23.1)
23.1
INTRODUCTION
455
Figure 23.1 Relationship between the geocentric and local geodetic coordinate
systems.
In Equation (23.1), changes in the local geodetic coordinate system and
geocentric coordinate system are represented by (⌬n,⌬e,⌬u) and (⌬X,⌬Y,⌬Z),
respectively. Multiplying the rotation matrices given in Section 18.7 and simplifying yields
⌬n
⫺⌬e
⌬u
⫽
⫺sin cos
⫺sin
cos cos
⫺sin sin
cos
cos sin
cos
0
sin
⌬X
⌬Y
⌬Z
(23.2)
The changes in the coordinates of the local geodetic system can be determined from the observation of azimuth Az, slant distance s, and vertical angle
v. As shown in Figure 23.2, the changes in the local geodetic coordinate
system can be computed as
⌬n ⫽ s cos v cos Az
⌬e ⫽ s cos v sin Az
(23.3)
⌬u ⫽ s sin v
From Figure 23.2, the following inverse relationships can also be
developed:
s ⫽ 兹⌬n2 ⫹ ⌬e2 ⫹ ⌬u2
Az ⫽ tan⫺1
v ⫽ sin⫺1
⌬e
⌬n
⌬u
s
(23.4)
456
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
Figure 23.2 Reduction of observations in a local geodetic coordinate system.
By combining Equations (23.2) and (23.4), the reduced observations can
be computed using changes in the geocentric coordinates. The resulting equations are
IJ ⫽ 兹⌬X 2 ⫹ ⌬Y 2 ⫹ ⌬Z 2
Azij ⫽ tan⫺1
⫺⌬X sin i ⫹ ⌬Y cos i
⫺⌬X sin i cos i ⫺ ⌬Y sin i sin i ⫺ ⌬Z cos i
vij ⫽ sin⫺1
⌬X cos i cos i ⫹ ⌬Y cos i sin i ⫹ ⌬Z sin i
兹⌬X 2 ⫹ ⌬Y 2 ⫹ ⌬Z 2
(23.5)
(23.6)
(23.7)
In Equations (23.5) to (23.7), i and i are the latitude and longitude of the
observing station Pi, ⌬X is Xj ⫺ Xi, ⌬Y is Yj ⫺ Yi, and ⌬Z is Zj ⫺ Zi. For
completeness, the equation for the zenith angle is
zij ⫽ cos⫺1
⌬X cos i cos i ⫹ ⌬Y cos i sin i ⫹ ⌬Z sin i
兹⌬X 2 ⫹ ⌬Y 2 ⫹ ⌬Z 2
(23.8)
Furthermore, since an angle is the difference between two azimuths, Equation
(23.6) can be applied to horizontal angles.
23.2
LINEARIZATION OF EQUATIONS
Equations (23.5) to (23.8) can be linearized with respect to the local geodetic
coordinates. The development of these equations is covered in Vincenty
(1989) and Leick (2004). The development of two coefficients are demonstrated and the final linearized prototype equations for slant distances, azimuths, horizontal angles, and vertical angles are listed in the following
23.2
457
LINEARIZATION OF EQUATIONS
subsections. It is important to note that the residuals exist implicitly in each
equation in order to make the equations consistent when using observations
containing random errors.
23.2.1
Slant Distance Observations
In the three-dimensional model, slant distance (also known as slope distance)
does not need to be reduced to either the station ground marks or ellipsoid
in the functional model. However, if they are not reduced, the values computed from Equations (23.5) to (23.8) are based on the instrument and reflector locations and not on the station’s ground marks. The prototype
equation for the slant distances observed is
a1 dni ⫹ a2 dei ⫹ a3 dui ⫹ a4 dnj ⫹ a5 dej ⫹ a6 duj ⫽ sij ⫺ IJ0
(23.9)
In Equation (23.9), the coefficients a1 to a6 are defined in Table 23.1, IJ0 is
the slant distance computed using Equation (23.5), and sij is the slant distance
observed.
To demonstrate the derivation of the coefficients, a1 is derived by taking
the partial derivative of the slant distance formula in Equation (23.4) with
respect to ni:
⭸s
⌬n
⫽⫺
⭸ni
s
(a)
By substituting ⌬n from Equation (23.3), the resulting equation for a1 is
⭸s
⫽ ⫺cos vij cos Azij
⭸ni
(b)
Following similar procedures, the remaining coefficients for Equation (23.9)
are derived.
23.2.2
Azimuth Observations
The prototype equation for the observed azimuths is
b1 dni ⫹ b2 dei ⫹ b3 dui ⫹ b4 dnj ⫹ b5 dej ⫹ b6 duj ⫽ ␣ ⫺ Az0
(23.10)
In Equation (23.10), ␣ is the observed azimuth and Az0 is its computed value
based on Equation (23.6) and approximate values for the station coordinates.
As an example, the coefficient for b1 is computed by taking the partial
derivative of the azimuth formula in Equation (23.4) with respect to ni:
458
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
TABLE 23.1 Coefficients for Linearized Equations in Equations (23.9) to
(23.11)
(a) a1 ⫽ ⫺(cos vij cos Azij)0
(b) a2 ⫽ ⫺(cos vij sin Azij)0
(c) a3 ⫽ ⫺(sin vij)0
(d ) a4 ⫽ ⫺(cos vji cos Azji)0
(e) a5 ⫽ ⫺(cos vji sin Azji)0
(f) a6 ⫽ ⫺(sin vji)0
(g) b1 ⫽
冉
再
再
再
冊
冋
sin Azij
IJ cos vij
冉
(h) b2 ⫽ ⫺
0
冊
cos Azij
IJ cos vij
(i) b3 ⫽ 0
0
册冎
sin Azij
sin j sin(j ⫺ i)
cos(j ⫺ i) ⫹
IJ cos vij
tan Azij
(j) b4 ⫽ ⫺
0
冎
cos Azij
(k) b5 ⫽
[cos(j ⫺ i) ⫺ sin i sin(j ⫺ i) tan Azij]
IJ cos vij
(l) b6 ⫽
0
cos Azij cos j
[sin(j ⫺ i) ⫹ (sin i cos(j ⫺ i)
IJ cos vij
冎
⫺ cos i tan j) tan Azij]
冉
冋
冋
冋
(m) c1 ⫽
(p) c4 ⫽
(q) c5 ⫽
(r) c6 ⫽
冊
sin vij cos Azij
IJ
0
(n) c2 ⫽
0
冉
冊
sin vij sin Azij
IJ
冉 冊
册
(o) c3 ⫽ ⫺
0
cos vij
IJ
0
⫺cos i sin j cos(j ⫺ i) ⫹ sin i cos j ⫹ sin vij cos vji cos Azji
IJ cos vij
册
0
⫺cos i sin(j ⫺ i) ⫹ sin vij cos vji sin Azji
IJ cos vij
0
册
sin vij sin vji ⫹ sin i sin j ⫹ cos i cos j cos(j ⫺ i)
IJ cos vij
0
⭸Azij
⌬n2
⌬e
⌬e
⫽
⫽
2
2
2
2
⭸ni
⌬n ⫹ ⌬e ⌬n
⌬n ⫹ ⌬e2
(c)
By substituting the appropriate formulas from Equation (23.3), the resulting
equation for b1 is
⭸Azij
s cos vij sin Azij
⫽ 2
2
⭸ni
s cos vij cos2 Azij ⫹ s2 cos2 vij sin2 Azij
⫽
s cos vij sin Azij
s2 cos2 vij (cos2 Azij ⫹ sin2 Azij)
⫽
sin Azij
s cos vij
(d)
23.2
LINEARIZATION OF EQUATIONS
459
The remaining coefficients of Equation (23.10) can be derived similarly.
For the remaining linearized observation equations in this section, the derivation of the coefficients for each equation follows procedures similar to those
presented in Equations (a) through (d). The derivations are left to the reader.
23.2.3
Vertical Angle Observations
Vertical angles are measured in the vertical plane and have a value of zero at
the horizon. As discussed in Section 23.7, vertical angles can be subject to
substantial systematic errors caused by deflection of the vertical and refraction. Due to these errors, vertical angles should not be used in an adjustment
on a regular basis. If these observations must be used in an adjustment, it is
important either to correct the observations for the systematic errors or to add
the unknown parameters to the mathematical model to correct for the systematic errors. As discussed in Section 23.7, adding correction parameters to a
typical survey runs the risk of overparameterization. This occurs when there
are more unknowns at a particular station than there are observations. In this
case the system is unsolvable. Thus, it is assumed that the corrections will
be made before the adjustment and the appropriate prototype equation is
c1 dni ⫹ c2 dei ⫹ c3 dui ⫹ c4 dnj ⫹ c5 dej ⫹ c6 duj ⫽ vij ⫺ v0
(23.11)
The coefficients for Equation (23.11) are listed in Table 23.1. Their values
are evaluated using the approximate coordinate values, vij is the observed
vertical angle, and v0 is the vertical angle computed using Equation (23.5)
and approximate station coordinates.
23.2.4
Horizontal Angle Observations
As stated earlier, horizontal angles are the difference in two azimuths. That
is, bif is computed as Azif ⫺ Azib, where b is the backsight station, i the
instrument station, and ƒ the foresight station. The prototype equation for a
horizontal angle is
d1 dnb ⫹ d2 deb ⫹ d3 dub ⫹ d4 dni ⫹ d5 dei ⫹ d6 dui ⫹ d7 dnƒ
⫹ d8 deƒ ⫹ d9 duƒ ⫽ bif ⫺ 0
(23.12)
The coefficients d1 through d9 in Equation (23.12) are listed in Table 23.2,
bif is the horizontal angle observed, and 0 is the value computed for the
angle based on difference between the foresight and backsight azimuths computed using Equation (23.6).
460
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
TABLE 23.2 Coefficients for Linearized Equation (23.12)
(s) d1 ⫽
再
冋
册冎
sin Azib
sin b sin(b ⫺ i)
cos(b ⫺ i) ⫹
IB cos vib
tan Azib
再
再
冎
0
cos Azib
[cos(b ⫺ i) ⫹ sin i sin(b ⫺ i) tan Azib]
IB cos vib
(t) d2 ⫽ ⫺
cos Azib cos b
(u) d3 ⫽ ⫺
[sin(b ⫺ i)
IB cos vib
冎
0
⫹ (sin i cos(b ⫺ i) ⫺ cos i tan b) tan Azib]
(v) d4 ⫽
冉
冊
sin Azif
sin Azib
⫺
IF cos vif IB cos vib
再
(y) d7 ⫽ ⫺
再
再
冋
(w) d5 ⫽
0
冉
0
冊
cos Azif
cos Azib
⫺
IB cos vib IF cos vif
册冎
sin Azif
sin ƒ sin(ƒ ⫺ i)
cos(ƒ ⫺ i) ⫹
IF cos vif
tan Azif
0
冎
cos Azif
(z) d8 ⫽
[cos(ƒ ⫺ i) ⫹ sin i sin(ƒ ⫺ i) tan Azif]
IF cos vif
(aa) d9 ⫽
cos Azif cos ƒ
[sin(ƒ ⫺ i)
IF cos vif
(x) d6 ⫽ 0
0
0
冎
⫹ (sin i cos(ƒ ⫺ i) ⫺ cos i tan ƒ) tan Azif]
0
23.2.5
Differential Leveling Observations
Orthometric height differences as derived from differential leveling also can
be included in the three-dimensional geodetic network adjustment model.
However, as discussed in Section 23.7, the inclusion of this observation type
requires a correction for geoidal height differences between the stations and
the application of orthometric corrections. Since the adjustment model is nonlinear, the observation equation for elevation differences as given in Equation
(11.1) must also be linearized. The prototype equation for differential leveling
between stations I and J is
1 duj ⫺ 1 dui ⫽ ⌬Hij ⫹ ⌬Nij ⫺ ⌬hij
(23.13)
In Equation (23.13), ⌬Hij is the elevation difference observed between the
stations, ⌬Nij the difference in geoidal height between the stations, and ⌬hij
the change in geodetic height.
23.2.6
Horizontal Distance Observations
As shown in Figure 23.3, the elevation differences of the endpoint stations
and converging radii cause the horizontal distances observed from opposite
23.2
LINEARIZATION OF EQUATIONS
461
Figure 23.3 Comparison of horizontal distances from opposite ends of the line.
ends of the lines to start and terminate at different points. Thus, technically,
l1 and l2 are not of the same length. In the local geodetic system, the observation equations for l1 and l2 are
l1 ⫽ 兹⌬n21 ⫹ ⌬e21
(23.14)
l2 ⫽ 兹⌬n ⫹ ⌬e
2
2
2
2
However, for the typically short distances observed in surveying practice,
l1 and l2, ⌬n1 and ⌬n2, and ⌬e1 and ⌬e2 are approximately equal. Letting l
equal l1 and l2, ⌬n equal ⌬n1 and ⫺⌬n2, and ⌬e equal ⌬e1 and ⫺⌬e2, the
linearized observation equation in local geodetic system is
冉
j ⫺ i
IJ
⫺
冉
冊
Mi
dni ⫹
0
i ⫺ j
IJ
冊
冉
i ⫺ j
IJ
冊
Nj dei ⫺
Nj dej ⫽ lij ⫺ l0
冉
j ⫺ i
IJ
冊
Mi
dnj
0
(23.15)
In Equation (23.15), the Mi is the radius in the meridian at the observing
station I, Ni the radius in the normal, lij the horizontal distance observed, and
l0 the distance computed using approximate coordinates and Equation (23.14).
The radii are computed as
462
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
Mi ⫽
a(1 ⫺ e2)
(1 ⫺ e2 sin2 i)3 / 2
(23.16)
a
Ni ⫽
兹1 ⫺ e2 sin2 i
23.3
MINIMUM NUMBER OF CONSTRAINTS
A three-dimensional geodetic network adjustment requires both horizontal
and vertical control. As discussed in Section 16.5, to fix the horizontal part
of the adjustment requires one station fixed in position and one line of known
direction. This can be accomplished by fixing the latitude and longitude of
one station along with the azimuth of a line or the longitude of a second
station. The vertical plane in the adjustment can be fixed with three benchmark stations. Since the adjustment is performed with geodetic heights, the
orthometric height of the benchmark stations must be corrected using Equation (23.37).
As discussed in Chapter 20, control can be adjusted or fixed simply by
setting the appropriate values in the stochastic model. In the case of vertical
control, benchmarks are often given as orthometric heights. Since the geoid
model is known only to about a centimeter, the standard deviations of geodetic
heights for benchmark stations should not be set any better than this value.
23.4
EXAMPLE ADJUSTMENT
To illustrate a three-dimensional least squares adjustment, the simple network
shown in Figure 23.4 will be used. The standard deviations for the coordinates
are shown in Table 23.3 in the local geodetic system. This system was chosen
Figure 23.4 Example of a three-dimensional geodetic network.
TABLE 23.3 Data for the Three-Dimensional Geodetic Network
Geodetic positions
Station
H (m)
N (m)
A
B
C
D
41⬚18⬘26.04850ⴖN
41⬚18⬘40.46660ⴖN
41⬚18⬘22.04010ⴖN
41⬚18⬘27.65860ⴖN
76⬚00⬘10.24860ⴖW
76⬚00⬘05.50180ⴖW
76⬚00⬘00.94390ⴖW
76⬚00⬘31.38550ⴖW
372.221
351.394
362.865
370.874
⫺31.723
⫺31.713
⫺31.726
⫺31.722
Station
Sn (m)
Se (m)
Sh (m)
A
B
C
0.001
0.001
—
—
—
—
0.01
0.01
0.01
Geodetic azimuth
Course
Azimuth
S (ⴖ)
AB
13⬚56⬘26.9ⴖ
0.001ⴖ
Course
Distance (m)
S (m)
AB
AC
CD
DA
BC
458.796
249.462
729.122
494.214
578.393
0.005
0.005
0.006
0.005
0.005
Slant distances
Horizontal angles
Stations (bif)
Angles
S (ⴖ)
DAB
BAC
CAD
ABC
CBA
BCD
98⬚10⬘25ⴖ
105⬚47⬘45ⴖ
156⬚01⬘44ⴖ
335⬚29⬘37ⴖ
24⬚30⬘19ⴖ
294⬚19⬘17ⴖ
2.8
3.5
4.1
2.4
2.4
2.3
Direction
Angle
S (ⴖ)
AC
CD
DA
AB
BC
92⬚09⬘01ⴖ
89⬚22⬘24ⴖ
89⬚50⬘44ⴖ
92⬚36⬘12ⴖ
88⬚52⬘01ⴖ
2.5
0.8
1.2
1.4
1.1
Zenith angles
Elevation differences
Stations
⌬Elev (m)
S (m)
AC
⫺9.359
0.005
463
464
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
because it is an intuitive system for assignment of realistic uncertainty values
in the adjustment and since the adjustment will be performed in the local
geodetic system. As shown in Table 23.3, the network is fixed in horizontal
position by weighting of the northing and easting coordinates of station A
and rotationally by overweighting the azimuth of line AB. The horizontal
rotation of the network could have also been fixed by weighting the easting
of station B.
The elevation datum was fixed in position and rotation by weighting the
height components of stations A, B, and C. Since orthometric heights were
given for the stations, the geoid separation was applied following Equation
(23.37) to compute geodetic heights for the stations. The systematic errors
discussed in Section 23.7 were removed previously from all angular observations. The results of the adjustment are shown in Figure 23.4.
Whereas the adjustment is performed in the local geodetic system, both
geodetic and geocentric coordinates are required to compute the coefficients
and computed observations. Since traditional observations are taken by an
elevated instrument to some elevated target, the geodetic heights of each station must be increased by the setup heights when computing geodetic coordinates. This simple addition to the software removes the need for reducing
observations to the station marks. In this example, the setup heights of the
instruments and targets are assumed to be zero and thus do not need to be
considered.
23.4.1
Addition of Slant Distances
Following prototype equation (23.9), each slant distance observation adds one
row to the system of equations. As an example for the slant distance AC, the
coefficients are computed as
⫺(cos vAB cos AzAB)0 dnA ⫺ (cos vAB sin AzAB)0 deA ⫺ (sin vAB)0 duA
⫺(cos vBA cos AzBA)0 dnB ⫺ (cos vBA sin AzBA)0 deB ⫺ (sin vBA)0 duB
⫽ sAB ⫺ AB0
(23.17)
In Equation (23.17) the values for AzAB and AzBA are computed using Equation (23.6), vAB and vBA are computed using Equation (23.7), and AB using
Equation (23.5). The numerical values, to five decimal places, for Equation
(23.17) are
⫺0.96954dnA ⫺ 0.24067deA ⫹ 0.04541duA ⫺ 0.96954dnB
⫹ 0.24069deB ⫺ 0.04534duB ⫽ 0.0017
(23.18)
23.4
465
EXAMPLE ADJUSTMENT
For each slant distance, a similar observation equation is written. The
reader should note in Equation (23.18) that unlike plane adjustments as presented in Chapter 14, the coefficients of the occupied and sighted stations
vary slightly due to Earth curvature. Four more equations for distances AC,
CD, DA, and BC are added to the system of equations.
23.4.2
Addition of Horizontal Angles
The angles observed in the network were corrected for the systematic errors
caused by the height of targets and deflection of the vertical as given in
Equations (23.30) and (23.32). Following prototype equation (23.12), an observation equation is written for each horizontal angle. As an example, the
observation equation for angle DAB is
再
冉
冊冎
sin AzAD
sin D sin(D ⫺ A)
cos(D ⫺ A) ⫹
AD cos vAD
tan AzAD
⫺
⫺
再
再
⫹
⫺
⫹
⫹
冎
cos AzAD
[cos(D ⫺ A) ⫹ sin A sin(D ⫺ A) tan AzAD]
AD cos vAD
deD
0
sin AzAD cos D
[sin(D ⫺ A) ⫹ (sin A cos(D ⫺ A)
AD cos vAD
冎
⫺ cos A tan D) tan AzAD]
⫹
dnD
0
冉
冉
再
再
再
duD
0
冊
冊
sin AzAB
sin AzAD
⫺
AB cos vAB AD cos vAD
0
cos AzAD
cos AzAB
⫺
AD cos vAD AB cos vAB
0
冋
dnA
deA ⫹ 0duA
册冎
sin AzAB
sin B sin(B ⫺ A)
cos(B ⫺ A) ⫹
AB cos vAB
tan AzAB
0
dnB
冎
cos AzAB
[cos(B ⫺ A) ⫹ sin A sin(B ⫺ A) tan AzAB]
AB cos vAB
deB
0
sin AzAB cos B
[sin(B ⫺ A) ⫹ (sin A cos(B ⫺ A)
AB cos vAB
冎
⫺ cos A tan B) tan AzAB]
0
duB ⫽ DAB ⫺ 0
(23.19)
466
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
In Equation (23.19) the values in the braces are evaluated at their approximate
coordinate values, DAB is the observed angular value corrected for the systematic errors discussed in Section 23.7, and 0 is the value computed based
on the difference in the computed values for the backsight (AD) and foresight
(AB) azimuths. Substituting the appropriate values into Equation (23.19) and
converting the radian values to units of s/m results in
⫺415.247dnD ⫺ 41.991deD ⫺ 0.00001duD ⫹ 523.669dnA ⫺ 394.828deA
⫹ 0duA ⫺ 108.432dnB ⫹ 436.790deB ⫺ 0.00003duB ⫽ 11.852ⴖ
(23.20)
For each angle, a similar observation equation is written, resulting in a total
of 10 equations to be added to the adjustment.
23.4.3
Addition of Zenith Angles
All the zenith angles in this example problem were corrected for deflection
of the vertical, refraction, and target height as discussed in Section 23.7.
Converting the zenith angle from A to C to an equivalent vertical angle and
following Equation (23.11), the observation equation is
冉
冊
sin vAC cos AzAC
AC
⫹
⫹
⫹
冢
冉
冉
0
dnA ⫹
冉
冊
sin vAC sin AzAC
AC
deA ⫺
0
冣
⫺cos A sin C cos(C ⫺ A) ⫹ sin A cos C
⫹ sin vAC cos vCA cos AzCA
AC cos vAC
0
冉
冊
cos vAC
AC
duA
0
dnC
冊
⫺cos A sin (C ⫺ A) ⫹ sin vAC cos vCA sin AzCA
AC cos vAC
0
deC
冊
sin vAC sin vCA ⫹ sin A sin C ⫹ cos A cos C cos(C ⫺ A)
AC cos vAC
⫽ vAC ⫺ v0
duC
0
(23.21)
In Equation (23.21) the coefficients on the left side of the equations are
evaluated using the approximate values of the coordinates in Equations (23.5)
to (23.7). The numerical values for Equation (23.21) in units of s/m are
15.395dnA ⫺ 26.947deA ⫺ 826.229duA ⫺ 15.379dnC
⫹ 26.919deC ⫹ 826.230duC ⫽ 2.971ⴖ
For each zenith angle, a similar observation equation is written, resulting in
a total of five equations to be added to the adjustment.
23.4
23.4.4
467
EXAMPLE ADJUSTMENT
Addition of Observed Azimuths
As discussed in Section 23.7, observed azimuths must be corrected for deflection of the vertical and height of the target. In this example, the azimuth
was already given as its geodetic value. Thus, the azimuth was used to fix
the adjustment by weighting. The observation equation for the azimuth is
冉
冊
sin AzAB
AB cos vB
⫺
⫹
⫹
再
再
再
dnA ⫺
0
冉
冊
cos AzAB
AB cos vAB
deA ⫹ 0duA
0
冎
冎
sin AzAB
[cos(B ⫺ A) ⫹ sin B sin(B ⫺ A) cos AzB]
AB cos vAB
0
cos AzAB
[cos(B ⫺ A) ⫺ sin A sin(B ⫺ A) tan AzAB]
AB cos vAB
0
dnB
deB
cos AzAB cos B
[sin(B ⫺ A) ⫺ (sin A cos(B ⫺ A)
AB cos vAB
冎
⫺ cos A tan B) tan AzAB]
duB
0
⫽ ␣AB ⫺ AzAB
(23.22)
In Equation (23.22) the coefficients on the left side are evaluated from the
approximate coordinate values using Equations (23.5) to (23.7). The numerical values of Equation (23.22), in units of s/m, are
108.42dnA ⫺ 436.79deA ⫹ 0duA ⫺ 108.43dnB
⫹ 436.79deB ⫺ 0.00003duB ⫽ ⫺0.256ⴖ
23.4.5
Addition of Elevation Differences
Leveling is a process of determining heights above the geoid. The data for
this example lists the orthometric and geoidal heights for each station. The
difference in orthometric heights requires the application of Equation (23.38)
to obtain the geodetic height differences between the stations. The observation
equation for the elevation difference between stations A and C is
1duC ⫺ 1duA ⫽ ⌬HAC ⫹ ⌬NAC ⫺ ⌬hAC
(23.23)
The resulting numerical values for Equation (23.23) are
1duC ⫺ 1duA ⫽ ⫺9.359 ⫹ (⫺31.726 ⫹ 31.723)
⫺ (331.139 ⫺ 340.498) ⫽ ⫺0.003 m
468
23.4.6
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
Adjustment of Control Stations
With the addition of the u coordinate, observation equations similar to those
presented in Section 20.2 can be added for each control coordinate. As stated
in Section 23.1, the local geodetic coordinate system has its origin at each
instrument station. Thus it is not a true coordinate system with a single origin.
Because of this, the approximate coordinates for each station must be stored
in either the geocentric or geodetic coordinate system. In this discussion, it
is assumed that the values are stored in the geodetic coordinate system. Again,
nonlinear versions of the equations must be written to match the overall nonlinear nature of the adjustment. The observation equations are
1dn ⫽ N ⫺ n
(23.24)
1de ⫽ E ⫺ e
(23.25)
1du ⫽ U ⫺ u
(23.26)
where (N,E,U) are the given control coordinate values and (n,e,u) are their
adjusted values. In the first iteration, the control coordinate values and their
adjusted values will be the same. In subsequent iterations, small variations
between the control coordinate values and their adjusted counterparts will be
observed. The observation equations for station A in the first iteration are
1dn ⫽ 0
1de ⫽ 0
1du ⫽ 0
Since the control coordinate values and approximate station coordinate
values will be expressed in terms of geodetic coordinates, the initial K-matrix
values given in Equations (23.24) to (23.26) will be expressed as changes in
geodetic coordinates. These values must then be transformed into the local
geodetic coordinate system. The relationships between changes in the two
systems are
dn
M⫹h
de ⫽
0
du
0
0
0
(N ⫹ h) cos 0
0
1
d
d
d ⫽ RLG d
dh
dh
(23.27)
In Equation (23.27), M represents the radius in the meridian and N the
radius in the normal at a station with latitude as given by Equation (23.16),
h the geodetic height, and RLG the transformation matrix between changes in
the geodetic and local geodetic coordinate systems. The K-matrix values ex-
23.4
469
EXAMPLE ADJUSTMENT
pressed in terms of geodetic coordinate system can be transformed into Kmatrix values of the local geodetic coordinate system using Equation (23.27).
23.4.7
Results of Adjustment
A partial list of the results of the adjustment from the ADJUST software
package is given in Figure 23.5. The solution converged in two iterations.
Note that the latitude and longitude of station A were held fixed, as was the
azimuth of line AB.
23.4.8
Updating Geodetic Coordinates
At the completion of each iteration, the corrections of dn, de, and du will be
determined for each station in the adjustment. However, geodetic coordinates
are used to represent station positions. Thus, after each iteration the local
geodetic coordinate system corrections of dn, de, and du must be transformed
into changes in the geodetic system using the inverse relationship of Equation
(23.27), or
di ⫽
dni
Mi ⫹ hi
di ⫽
dei
(Ni ⫹ hi) cos i
dhi ⫽ dei
(23.28)
In Equation (23.28), the corrections to the latitude, longitude, and geodetic
height of station I are di, di, and dhi, respectively. All other terms are as
defined previously. Similarly, the uncertainties for each station will be in the
local geodetic system. Although these uncertainties can be used to represent
northing, easting, and geodetic height errors at each station, Equation (6.13)
must be used to transform these uncertainties into the geodetic system as
,,h
⫽ R⫺1
n,e,u
R⫺T
(23.29)
where
,,h
n,e,u
⫽
⫽
,
2
,h
,h
,h
2h
n,e
2e
e,u
n,u
e,u
2u
2
,
,h
2n
n,e
n,u
R⫽
M⫹h
0
0
0
(N ⫹ h) cos 0
0
0
1
Note that each station’s uncertainties can be computed using the 3 ⫻ 3 block
diagonal elements from the Qxx matrix of the adjustment.
470
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
Adjusted Geodetic Coordinates
Station
Latitude
Longitude
Orth height
N
S-Lat(ⴖ)
S-Lon(ⴖ)
S-h
=============================================================================================
A
41⬚1826.04850ⴖN
76⬚0010.24860ⴖW
372.2232
⫺31.72
0.000065
0.000087
0.0119
B
41⬚1840.46653ⴖN
76⬚0005.50185ⴖW
351.3904
⫺31.71
0.000231
0.000113
0.0120
C
41⬚1822.04015ⴖN
76⬚0000.94395ⴖW
362.8664
⫺31.73
0.000162
0.000270
0.0119
D
41⬚1827.65788ⴖN
76⬚0031.38547ⴖW
370.8748
⫺31.72
0.000304
0.000353
0.0124
Station Statistics
Station
Sn
Se
t-ang⬚
Su
Sv
S-Elev#
=============================================================================================
A
0.0020
0.0020
45.0000
0.0020
0.0020
0.0119
B
0.0071
0.0026
13.9417
0.0073
0.0020
0.0120
C
0.0050
0.0063
96.0029
0.0063
0.0050
0.0119
D
0.0094
0.0082
175.3588
0.0094
0.0082
0.0124
#-Values do not include uncertainty in geoidal height.
Adjusted Slope Distances
Station
Station
Occupied
Sighted
Distance
V
=============================================================================================
A
B
458.792
⫺0.004
A
C
249.468
0.006
C
D
729.116
⫺0.006
D
A
494.213
⫺0.001
B
C
578.399
0.006
Adjusted Mark-to-Mark and Geodetic Distances
Station
Station
Mark-to-Mark
Geodetic
Occupied
Sighted
Distance
Distance
=============================================================================================
A
B
458.792
458.317
A
C
249.468
249.282
C
D
729.116
729.034
D
A
494.213
494.182
B
C
578.399
578.255
Adjusted Angle Observations
Station
Station
Station
Backsighted
Occupied
Foresighted
Angle
V(ⴖ)
=============================================================================================
D
A
B
98⬚1022.13ⴖ
⫺2.87
B
A
C
105⬚4752.54ⴖ
7.54
C
A
D
156⬚0145.34ⴖ
1.34
A
B
C
335⬚2934.52ⴖ
⫺2.48
C
B
A
24⬚3025.48ⴖ
6.48
B
C
D
294⬚1909.93ⴖ
⫺7.07
D
C
A
15⬚5908.09ⴖ
7.09
A
C
B
49⬚4141.98ⴖ
⫺3.02
C
D
A
352⬚0053.43ⴖ
8.43
A
D
C
7⬚5906.57ⴖ
⫺5.43
Adjusted Azimuth Observations
Station
Station
Occupied
Sighted
Azimuth
V
=============================================================================================
A
B
13⬚5626.9ⴖ
0.00ⴖ
Figure 23.5 ADJUST listing of adjustment results for the example problem of Figure
23.4.
23.6
COMMENTS ON SYSTEMATIC ERRORS
471
Adjusted Zenith Angle Observations
Station
Station
Occupied
Sighted
Zenith Angle
V(ⴖ)
=============================================================================================
A
B
92⬚3612.2ⴖ
⫺0.19ⴖ
A
C
92⬚0904.7ⴖ
⫺3.69ⴖ
C
D
89⬚2225.1ⴖ
⫺1.05ⴖ
D
A
89⬚5045.6ⴖ
⫺1.61ⴖ
B
C
88⬚5201.3ⴖ
⫺0.26ⴖ
Adjusted Elevation Difference Observations
Station
Station
Elevation
Occupied
Sighted
Diff.
V
=============================================================================================
A
C
⫺9.363
⫺0.0038
Iterations
Redundancies
Reference Variance
Reference So
⫽
⫽
⫽
⫽
2
15
4.075
2.019
Figure 23.5 (Continued )
23.5
BUILDING AN ADJUSTMENT
Since this is a nonlinear adjustment, initial approximations are required for
all station parameters. Very good initial values for horizontal coordinates can
be determined using the procedures outlined in Chapter 16, with horizontal
observations reduced to a map projection surface. Following the adjustment,
grid coordinates can be converted to their geodetic equivalents. Similarly,
orthometric heights of stations can be determined using the procedures discussed in Chapter 12. The resulting orthometric heights can be converted to
geodetic heights using Equation (23.37). Since these are initial approximations, it is possible to use an average geoidal height N for the region when
applying Equation (23.37). However, this procedure will result in more uncertainty in the resulting heights.
Before combining a large set of three-dimensional conventional observations, it is wise to perform adjustments using the horizontal and vertical functional models presented in Chapters 12 and 16 to isolate potential blunders.
For large regional data sets, a smaller data subset should be adjusted to isolate
blunders. After these smaller data sets are cleaned and adjusted, a combined
adjustment can be attempted with better initial approximations for the unknowns and the knowledge that most if not all blunders have been removed
from the data.
23.6
COMMENTS ON SYSTEMATIC ERRORS
Usually, small local surveys result in systematic error components that are
small enough to be considered negligible. However as the size of the survey
472
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
increases, variations in the direction of plumb lines must be taken into account. This error can be removed from angular observations prior to the adjustment with a priori knowledge of the deflection components and at
each station. Modeled values for these components can be obtained using
software such as DEFLEC99,1 which is available from the National Geodetic
Survey. The systematic error corrections to the zenith angles, azimuths, and
horizontal angles are
zij ⫽ zij ⫹
Azij ⫽ Azobs ⫹
i
i
cos Azij ⫹
i
sin Azij
(23.30)
tan i ⫹ (i sin Azobs ⫺ i cos Azobs) cot zij
(23.31)
∠BIF ⫽ ∠BIFobs ⫹ (i sin Azif ⫺ i cos Azif) cot zif
⫺ (i sin Azib ⫺ i cos Azib) cot zib
(23.32)
where i and i are the deflection of the vertical components at the observation
station I; zij⬘ is the zenith angle observed from station I to J; zij the corrected
zenith angle between the same stations; Azobs the astronomical azimuth observed between station I and J; Azij the corrected azimuth; ∠BIFobs the angle
observed where station B is the backsight station, I the instrument station,
and F the foresight station; and ∠BIF the corrected horizontal angle.
Additionally, target height differences must also be considered for both
directions and angles. The azimuth of a line corrected for target height is
Azc ⫽ ␣ ⫹
0.108ⴖ cos2 i sin 2␣ h
1000
(23.33)
In Equation (23.33), Azc is the corrected azimuth, ␣ the observed azimuth,
i the latitude of the observation station, and h the geodetic height of the
target in meters. Since an angle is simply the difference between the foresight
and backsight azimuths, the correction to an angle due to the height of the
targets is
冋
c ⫽ bif ⫹ 0.108ⴖ cos2 i
册
hƒ sin 2␣if sin 2␣ib hb
⫺
1000
1000
(23.34)
where c is the corrected angle, bif the observed angle, and ␣if and ␣ib the
azimuths of the foresight and backsight lines, respectively.
Atmospheric refraction must be considered when including vertical or zenith angles in an adjustment. Since the correction for atmospheric refraction
1
DEFLEC99 is available at http: / / www.ngs.noaa.gov / PC PROD / pc prod.shtml#DEFLEC99.
23.6
COMMENTS ON SYSTEMATIC ERRORS
473
is so difficult to model, one can only hope to account for a portion of correction in an adjustment. The remainder of the correction must be modeled
in the adjustment or be absorbed in the residuals of the observation. The
correction of the first-order effect of refraction can be determined by observing simultaneously vertical or zenith angles at each end of the line IJ. From
these observations, the first-order correction ⌬Zij for reciprocal zenith angles
is determined as
⌬Zij ⬅ ⌬Zji ⫽ 90⬚ ⫺ [0.5(Zij ⫹ Zji) ⫺ ]
(23.35)
In Equation (23.35), Zij and Zji are the reciprocal zenith angles observed simultaneously and is the correction for the deflection of vertical, computed
as
⫽ ⫺(i cos Azij ⫹ i sin Azij)
(23.36)
It is possible to create an adjustment model that includes correction terms
for , , and atmospheric refraction. However, due to the limited number of
observations in a typical survey network, there is a danger of overparameterization (Leick, 2004). Thus, it is recommended that these corrections be applied to the angles before carrying out the adjustment.
Unfortunately, to determine the deflection of the vertical components (
and ), the latitude and longitude of the station must be known before completion of the adjustment. Approximate values for latitude and longitude will
suffice in most instances. However, in instances where good approximations
for the geodetic coordinates are not known, it is possible to correct the original
observations after an initial adjustment to ensure that the geodetic coordinates
for each station are close to their final values. In this case, after correcting
the observations, the adjustment could be run a second time.
A similar problem exists with differential leveling observations. Geodetic
height h and orthometric height H differ by the geoidal height N, or
h⫽H⫹N
(23.37)
Elevation differences as determined by differential leveling must account
for the differences in geoidal height between the benchmark stations. This
situation can be described mathematically as
hj ⫺ hi ⫽ (Hj ⫹ Nj) ⫺ (Hi ⫹ Ni)
⌬h ⫽ ⌬H ⫹ ⌬N
(23.38)
The correction to leveled height differences should be performed when
these observations are included in the three-dimensional geodetic network
474
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
adjustment. Values for geoidal height can be determined using the GEOID03
software available from the National Geodetic Survey.2 Similar to deflection
of the vertical parameters, it is possible to include geoidal height parameters
in the mathematical model of the adjustment. Again, this approach runs a risk
of overparameterization in the adjustment.
The orthometric correction OC for a leveling line can be approximated
using a formula derived by Bomford (1980) as
OC ⫽
⫺H(⌬ⴖ)(0.0053 sin 2)
(23.39)
where H is the orthometric height at the instrument station in meters, ⌬ⴖ the
change in latitude between the backsight and foresight stations in units of
seconds, the latitude of the instrument station, and the conversion from
seconds to radians, which is approximately 206,264.8ⴖ /rad. Obviously, Equation (23.39) would be tedious to apply for a substantial north–south leveling
circuit. However, for most typical surveys involving small regions, this correction can be very small and easily absorbed in the residuals of the adjustment. For instance, a 120-km north–south leveling line at approximately 42⬚
latitude would result in an orthometric correction of about 2 cm.
PROBLEMS
Note: Unless otherwise specified, use the GRS 80 ellipsoidal parameters of a
⫽ 6.378.137.0 m and ƒ ⫽ 1/298.257222101 for the following problems.
23.1
Using the data supplied in Figure 23.4, what value is computed for
the slant distance AB?
23.2
Repeat Problem 23.1 for slant distance AC.
23.3
Using the data supplied in Figure 23.4, what value is computed for
the azimuth AB?
23.4
Using the data supplied in Figure 23.4, what value is computed for
the horizontal angle DAB?
23.5
Repeat Problem 23.4 for horizontal angle ACB.
23.6
Using the data supplied in Figure 23.4, what value is computed for
vAB?
2
GEOID03 software is available at http: / / www.ngs.noaa.gov.
PROBLEMS
475
23.7
Repeat Problem 23.6 for the vertical angle vAC.
23.8
Develop the observation equation for slant distance AC in Figure 23.4.
23.9
Repeat Problem 23.8 for slant distance CD.
23.10 Develop the observation equation for horizontal angle BAC.
23.11 Repeat Problem 23.10 for horizontal angle ACB.
23.12 Develop the observation equation for zenith angle AB.
23.13 The approximate values for the geodetic coordinates of the endpoints
of line AB are A: (39⬚1843.2000ⴖ N, 106⬚4640.8000ⴖ W, 1191.161
m) and B: (39⬚1838.5210ⴖ N, 106⬚4622.6822ⴖ W, 1217.239 m). Using the WGS 84 ellipsoid, determine the computed:
(a) slant distance AB.
(b) azimuth AB.
(c) vertical angle AB.
23.14 Repeat Problem 23.13 with approximate values for the geodetic coordinates of A: (36⬚4452.8000ⴖ N, 119⬚4615.6000ⴖ W, 90.221 m)
and B: (36⬚4504.9159ⴖ N, 119⬚4617.9521ⴖ W, 89.675 m).
23.15 Develop the observation equation for slant distance AB if the distance
observed in Problem 23.13 is 458.237 m.
23.16 Develop the observation equation if the distance observed in Problem
23.14 is 378.011 m.
23.17 Develop the azimuth observation equation for Problem 23.13 if the
azimuth observation is 108⬚2319ⴖ. (Assume that all systematic errors
are removed from the observation.)
23.18 Develop the azimuth observation equation for Problem 23.14 if the
azimuth observation is 351⬚0724ⴖ. (Assume that all systematic errors
are removed from the observation.)
Use the following data and the GRS 80 ellipsoid in Problems 23.19 to 23.26.
Geodetic coordinates
Station
Latitude
Longitude
A
B
C
D
40⬚2528.7700ⴖN
40⬚2546.3978ⴖN
40⬚2543.3596ⴖN
40⬚2533.6000ⴖN
86⬚5415.0464ⴖW
86⬚5412.3908ⴖW
86⬚5430.5172ⴖW
86⬚5428.8000ⴖW
Height (m) Sn (m) Se (m) Su (m)
92.314
80.004
87.268
95.585
0.001
—
—
—
0.001
—
—
—
0.01
0.01
0.01
—
476
THREE-DIMENSIONAL GEODETIC NETWORK ADJUSTMENT
Angle observations
B
I
F
Angle
S (ⴖ)
hb (m)
hi (m)
hf (m)
D
C
B
A
D
C
B
A
D
C
B
A
A
A
A
B
B
B
C
C
C
D
D
D
C
B
D
D
C
A
A
D
B
B
A
C
26⬚1807.6ⴖ
45⬚3506.7ⴖ
288⬚0642.4ⴖ
37⬚5041.9ⴖ
33⬚1303.8ⴖ
288⬚5609.2ⴖ
63⬚2053.4ⴖ
31⬚2136.2ⴖ
265⬚1726.4ⴖ
52⬚0425.0ⴖ
70⬚1556.7ⴖ
237⬚3941.8ⴖ
4.7
4.1
5.0
4.1
4.3
4.6
4.5
5.1
5.9
5.3
5.0
6.6
1.303
1.402
1.295
1.305
1.500
1.206
1.300
1.425
1.398
1.500
1.500
1.500
1.295
1.405
1.301
1.398
1.500
1.210
1.300
1.423
1.205
1.500
1.500
1.400
1.300
1.398
1.305
1.299
1.500
1.208
1.300
1.420
1.300
1.500
1.500
1.500
Slant distance observations
From
To
Distance (m)
S (m)
hi (m)
hr (m)
A
B
C
D
A
B
B
C
D
A
C
D
547.433
437.451
303.879
356.813
579.263
552.833
0.008
0.008
0.008
0.008
0.008
0.008
1.400
1.497
1.500
1.302
1.300
1.500
1.500
1.595
1.500
1.296
1.500
1.400
Azimuth observations
From
To
Azimuth
S (ⴖ)
hi (m)
hr (m)
A
B
6⬚3404.6ⴖ
0.001
1.500
1.500
Elevation differences
From
To
⌬Elev (m)
S (m)
A
B
C
D
⫺5.053
0.012
0.012
13.585
23.19 Use a computational program to develop the coefficient matrix for
the angular data in the table. List all coefficients in units of s/m.
23.20 Use a computational program to develop the coefficient matrix for
the slant distance data in the table.
23.21 Use a computational program to develop the observation equation for
the azimuth data in the table. List the coefficients in units of s/m.
23.22 Use a computational program to develop the observation equations
for the elevation differences in the table.
PROBLEMS
477
23.23 Use ADJUST to determine the most probable values for the geodetic
coordinates in the table using only the angle, azimuth, and slant distance observations.
23.24 Use ADJUST to determine the most probable values for the geodetic
coordinates in the table using all the observations. Remember to account for the geoidal height at each station.
23.25 Repeat Problem 23.24 by removing azimuth AB from the adjustment
and overweighting the easting of station B. The standard deviations
of station B should be Sn ⫽ 1000 m, Se ⫽ 0.001 m, and Su ⫽
0.01 m.
23.26 What are the adjusted orthometric heights of the stations in the table?
23.27 If ⫽ 4.87ⴖ and ⫽ 0.01ⴖ are the deflection of the vertical components at station A in Section 23.4, what is the correction to the zenith
angle going from station A to B?
23.28 Repeat Problem 23.27 for the zenith angle from station A to C.
23.29 If the zenith angle observed from A to B in Figure 23.6 is 91⬚0454ⴖ
and the deflection of the vertical components are ⫽ 1.28ⴖ and ⫽
3.58ⴖ, what is the corrected zenith angle?
23.30 Using the data given in Problem 23.27, what is the deflection of the
vertical correction for the azimuth AB?
23.31 Using the data given in Problem 23.29, what is the deflection of the
vertical correction for the azimuth AB?
23.32 Using the data given in Problem 23.27, what is the deflection of the
vertical correction for the angle DAB?
23.33 Using the data given in Problem 23.29, what is the deflection of the
vertical correction for the angle DAC?
23.34 Do Problem 23.33 using angle CAB.
Programming Problems
23.35 Develop a computational program that corrects vertical angles for
deflection of the vertical components.
23.36 Develop a computational program that computes the coefficients for
the azimuths, angles, slant distances, and elevation differences. Check
your program using data from Figure 23.6.
23.37 Modify the Mathcad worksheets provided for Chapter 23 on the CD
that accompanies this book to incorporate target and instrument
heights. Solve Problem 23.6.
CHAPTER 24
COMBINING GPS AND
TERRESTRIAL OBSERVATIONS
24.1
INTRODUCTION
Ellipsoids define the mathematical shape of Earth or a portion thereof. Ellipsoids are commonly defined by the length of their semimajor axis, a, and the
flattening factor, ƒ. Commonly used ellipsoids are the Geodetic Reference
System of 1980 (GRS 80) and the World Geodetic System of 1984 (WGS
84). Table 24.1 lists the length of the semimajor axis and flattening factor for
these ellipsoids.
A network of points determined with respect to an ellipsoid is a datum.
Datums define the geodetic coordinates of the points and thus the origin and
orientation of the datum. Most datums are regional in nature since the network
of points covers only a portion of the Earth. These are known as local datums.
For example, the North American Datum of 1983 (NAD 83) is a local datum
consisting of a network of points in Canada, the United States, Mexico, and
some Caribbean islands. However, the International Terrestrial Reference
Frame (ITRF) is an example of a datum defined by a multitude of points
located on all of Earth’s major land masses. This is known as a global datum.
Similarly, the coordinates of the GPS satellites are determined by a global
network of tracking stations on a unique datum. Stated incorrectly, GPS is
said to be using the WGS 84 datum. This statement should be interpreted by
the reader as the datum defined by the global network of tracking stations
that use the WGS 84 ellipsoid.
Besides their reference ellipsoid, datums differ in origin (translation), scale,
and rotations about the three cardinal axes. Many modern local datums and
global datums have nearly aligned coordinate axes and differ by only a few
meters in their origins. Thus, it is possible to use a three-dimensional coor478
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
24.1
INTRODUCTION
479
TABLE 24.1 Defining Ellipsoidal Parameters
Ellipsoid
a (m)
1/ƒ
GRS80
WGS84
6,378,137.0
6,378,137.0
298.257222101
298.257223563
dinate transformation to transform points from a local datum to a global
datum and back again. When using GPS and absolute positioning techniques,
the satellites serve as control points, and all points determined by this method
are defined in either the ITRF or WGS 84 datum, depending on the source
of ephemeris. However, this method of surveying is accurate only to the meter
level. Thus, relative positioning techniques are generally used. However, if
GPS receivers are placed on control points defined in a local datum such as
NAD 83, the resultant points are defined in a hybrid of the WGS 84 or ITRF
and the local datum. The differences in these systems are generally at the
centimeter level. For lower-order surveys, this difference may not be of much
importance. However, for higher-order control surveys, these differences must
be taken into account.
To combine GPS and terrestrial observations, the control from two different
datums must be reconciled. This can be done by transforming the local control
coordinates into the GPS reference datum, or by transforming the GPS established points into the local datum.
As discussed in Chapter 18, a three-dimensional conformal coordinate
transformation will take coordinates from one three-dimensional coordinate
system into another. However, since the datums are nearly aligned, the rotational angles are generally very small. As shown in Section 24.2, the rotational
process, as well as the entire transformation, can be simplified.
To perform the transformation, coordinates of common points in both systems must be placed into their respective geocentric coordinate systems (see
Section 17.4). Following this conversion, a least squares adjustment can be
performed to determine the transformation parameters between the two systems. Once the transformation parameters are determined, the coordinates of
any remaining points can be transformed.
For example, assume that GPS is being used to stake out a highway alignment and that the highway alignment was designed using control points from
the state plane coordinate system of 1983 (SPCS 83).1 The real-time GPS
datum is defined by a set of Department of Defense tracking stations using
the WGS 84 ellipsoid. The geodetic control used in the local state plane
coordinate system is based on a series of National Spatial Reference stations
in the United States, Canada, and Mexico using the GRS 80 ellipsoid. The
NEH coordinates of the highway design stations must be transformed from
1
Map projection coordinates systems are discussed in Appendix F.
480
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
the SPCS 83 coordinate system and orthometric heights to geodetic coordinates. The geodetic coordinates are then transformed to geocentric coordinates2 of (X,Y,Z)LS. These points must be occupied using a GPS receiver and
GPS coordinates (X,Y,Z)GPS derived from satellite observations. Once three
common points are occupied, the transformation parameters can be determined using a Helmert transformation. However, the prudent surveyor will
use four common points to increase the number of redundant measurements.
Following this, all GPS coordinates are transferred into the local coordinate
system, and stakeout of stations can be performed.
24.2
HELMERT TRANSFORMATION
Local datums such as NAD 83 are known as Earth-centered, Earth-fixed
(ECEF) coordinate systems. This means that the Z axis is nearly aligned with
the Conventional Terrestrial Pole, X axis with the Greenwich Meridian, and
the origin is at the mass center of the Earth as derived by the datum points
used in the definition. International datums such as the International Terrestrial Reference Frame use the same definitions for the axes, origin, and ellipsoid, but differ slightly due to the difference in the datum points used in its
determination. Thus, the rotational parameters and translations between two
ECEF coordinate systems are usually very small. The scale factor between
two datums using the same units of measure should be nearly 1.
Since the sine of a very small angle is equal to the angle in radians, the
cosine of a very small angle is nearly 1, and the product of two very small
numbers is nearly zero, for two nearly aligned coordinate systems Equation
(18.14) can be simplified as
1
R⬘ ⫽ ⫺3
2
3
1
⫺2
1
⫺1
1
⫽I⫹
0
⫺⌬3
⌬2
⌬3
0
⫺⌬1
⫺⌬2
⌬1
0
⫽ I ⫹ ⌬R
(24.1)
In Equation (24.1) ⌬1, ⌬2, and ⌬3 are in units of radians and have been
separated from the full matrix by the addition of the unit matrix. The introduction of ⌬ indicates that these values are differentially small.
The transformation of coordinates from one local datum to another datum
is performed as
XLD ⫽ sR⬘XGD ⫹ T
(24.2)
2
Equations to convert geodetic coordinates to geocentric coordinates are given in Section 17.5.
Appendix F has equations to convert geodetic coordinates to state plane coordinates.
24.2
481
HELMERT TRANSFORMATION
where s is the scale factor, XGD the (x,y,z) coordinates from the global data
(GD) to be transferred into the local datum (LD), and T the xyz translations
needed to coincide the origins of the two datums. Similarly, the scale factor,
s, and translation parameters, T, can be modified as
s ⫽ 1 ⫹ ⌬s
(24.3)
T ⫽ T0 ⫹ ⌬T
(24.4)
In Equation (24.4), the approximate shift vector, T0, can be computed as
x
T0 ⫽ y
z
⫺
LD
x
y
z
GD
⌬Tx
and ⌬T ⫽ ⌬Ty
⌬Tz
(24.5)
Since Equation (24.2) is nonlinear, a single common point or an average of
all the common points can be used in Equation (24.5) to obtain initial approximations. For a single station, i, the linearized model for the corrections
is
(24.6)
xLDi ⫺ xGDi ⫺ T0 ⫽ Ji dx
where
xi
0 ⫺zi
yi 1 0 0
Ji ⫽ yi
zi
0 ⫺xi 0 1 0
zi ⫺yi
xi
0 0 0 1
dx ⫽
⌬s
⌬1
⌬2
⌬3
⌬Tx
⌬Ty
⌬Tz
(24.7)
Example 24.1 The geocentric coordinates in the North American Datum of
1983 (NAD 83) and, from a GPS adjustment using a precise ephemeris, from
the International Terrestrial Reference Frame of 2000 (ITRF 00) datum are
shown below. Determine the transformation parameters to transform the additional station into the local NAD 83 datum.
NAD 83
ITRF 00
Station
x (m)
y (m)
z (m)
x (m)
y (m)
z (m)
A
E
F
G
B
1,160,604.924
1,160,083.830
1,160,786.583
1,160,648.090
⫺4,655,917.607
⫺4,655,634.217
⫺4,655,564.279
⫺4,656,191.149
4,188,338.994
4,188,722.346
4,188,637.038
4,188,040.857
1,160,374.046
1,159,852.942
1,160,555.699
1,160,417.220
1,160,406.372
⫺4,655,729.681
⫺4,655,446.265
⫺4,655,376.341
⫺4,656,003.238
⫺4,655,402.864
4,188,609.031
4,188,992.349
4,188,907.065
4,188,310.909
4,188,929.400
482
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
SOLUTION First iteration: Using station A, T0 is
1,160,604.924 ⫺ 1,160,374.046
230.877
T0 ⫽ ⫺4,655,917.607 ⫹ 4,655,729.681 ⫽ ⫺187.926
4,188,338.994 ⫺ 4,188,609.031
⫺270.037
The J and K matrices are
1160374.046
⫺4655729.681
4188609.031
1159852.942
⫺4655446.265
4188992.349
J⫽
1160555.699
⫺4655376.341
4188907.065
1160417.220
⫺4656003.238
41883120.909
0.000
4188609.031
4655729.681
0.000
4188992.349
4655446.265
0.000
4188907.065
4655376.341
0.000
4188310.909
4656003.238
⫺4188609.031
0.000
1160374.0464
⫺4188992.349
0.000
1159852.942
⫺4188907.065
0.000
1160555.699
⫺4188310.909
0.000
1160417.220
⫺4655729.681
⫺1160374.046
0.000
⫺4655446.265
⫺1159852.942
0.000
⫺4655376.341
⫺1160555.699
0.000
⫺4656003.238
⫺1160417.220
0.000
1
0
0
1
0
0
1
0
0
1
0
0
0
1
0
0
1
0
0
1
0
0
1
0
0
0
1
0
0
1
0
0
1
0
0
1
0.000
0.000
0.000
0.011
⫺0.026
0.034
K⫽
0.006
⫺0.012
0.010
⫺0.008
0.015
⫺0.015
Solving Equation (11.37) using the matrices above and adding X to the initial
values results in
X⫽
0.0000000256
⫺0.0000486262
⫺0.0000388234
⫺0.0000146664
⫺230.9285
186.7768
271.3325
s ⫽ 1 ⫹ 0.0000000256 ⫽ 1.0000000256
1 ⫽ ⫺0⬚00⬘10.02987ⴖ
2 ⫽ ⫺0⬚00⬘08.0079ⴖ
3 ⫽ ⫺0⬚00⬘03.02517ⴖ
Tx ⫽ 230.877 ⫺ 230.92 ⫽ ⫺0.052
Ty ⫽ ⫺187.926 ⫹ 186.7768 ⫽ ⫺1.149
Tz ⫽ ⫺270.037 ⫹ 271.3325 ⫽ 1.295
24.2
HELMERT TRANSFORMATION
483
The next iteration resulted in negligible changes in the translations, and
thus the transformation parameters are as listed. Using Equation (24.2), transformation parameters, and the GPS-derived ITRF 00 coordinates for station
B, the NAD 83 coordinates for station B are
geocentric coordinates ⫽ (1,160,637.257 ⫺4,655,590.805, 4,188,659.377)
geodetic coordinates ⫽ (41⬚18⬘40.46653ⴖ N, 76⬚00⬘05.50185ⴖ W,
319.677 m)
SPCS 83 coordinates ⫽ (746,397.796 m, 128,586.853 m)
The reader should note that the NAD 83 coordinates for station B are the
same as those listed in Figure 23.5. However, slight differences in coordinates
can be expected, due to random errors in the transformation.
Example 24.1 demonstrates the mathematical relationship between the two
datums for this small set of points. This process should always be considered
when combining traditional observations with GPS coordinates in higherorder surveys. That is, if GPS derived coordinates are to be entered into an
adjustment, they should first be transformed into the local datum. In the
United States, the National Geodetic Survey has developed horizontal timedependent positioning (HTDP) software3 that allows the users to transform
coordinates between NAD 83, WGS 84, and ITRF 00. This software also
takes into account plate tectonics by applying velocity vectors to stations
when they are known.
The process of transforming points from global datums to local datums is
important when performing RTK-GPS stakeout surveys. A GPS survey implicitly uses the points located in a global datum. Since engineering plans are
generally developed in a local datum such as NAD 83, the GPS-derived coordinate values must be transformed into the local coordinate system. This
process is known as localization (sometimes called site calibration) by manufacturers. From a design point of view, it is important to recognize that the
best results will be obtained if points common in both datums surround the
project area. After entering in the local datum coordinates, the GPS receiver
should occupy each station. The software will then compute the transforma-
3
A description of the HTDP software can be located on the NGS Web site at http: / / www.
ngs.noaa.gov / TOOLS / Htdp / Htdp.shtml. A Mathcad worksheet that demonstrates this software
is available on the CD that accompanies this book.
484
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
tion parameters and use these to determine the coordinate values for the points
to be surveyed.
24.3
ROTATIONS BETWEEN COORDINATE SYSTEMS
GPS uses the geocentric coordinate system and list its baseline vectors in this
system. Often, it is preferable to obtain the coordinate changes in terms of
geodetic coordinates. From geodesy, the relationship between changes in the
geodetic coordinate system and the geocentric coordinate system is
dX
dx ⫽ dY
dZ
⫽
⫺(M ⫹ h) sin sin
⫺(M ⫹ h) sin sin
(M ⫹ h) cos
⫺(N ⫹ h) cos sin
(N ⫹ h) cos cos
0
cos cos
cos sin
sin
d
⫽ RXG d
d
d
dh
(24.8)
dh
where dx represents the changes in the geocentric coordinates, M is the radius
in the meridian, and N the radius in the normal at latitude as given by
Equation (23.16), h is the geodetic height of the point, and RXG is the transformation matrix.
The transformation between changes in the geodetic coordinate system and
the local geodetic coordinate system are given in Equation (23.29), which is
repeated here for convenience:
dn
M⫹h
de ⫽
0
du
0
0
0
(N ⫹ h) cos 0
0
1
d
d
d ⫽ RLG d
dh
dh
(24.9)
where RLG is the rotation matrix between the geodetic and local geodetic
coordinate systems.
24.4 COMBINING GPS BASELINE VECTORS WITH
TRADITIONAL OBSERVATIONS
As discussed in Chapter 23, the three-dimensional geodetic adjustment allows
the adjustment of all traditional surveying observations. If the control is
known in the local coordinate system, Equation (24.2) can be included in
adjustment to account for datum differences.
24.4
COMBINING GPS BASELINE VECTORS WITH TRADITIONAL OBSERVATIONS
485
Baseline vectors are the geocentric coordinate differences between two
points. Thus, the translation component of Equation (24.2) is removed, leaving only the scaling and rotational parameters, and the last three columns of
the Ji matrix in Equations (24.6) and (24.7) are eliminated, leaving
xi
0 ⫺zi
yi
Ji ⫽ yi
zi
0 ⫺xi
zi ⫺yi
xi
0
dx ⫽
⌬s
⌬1
⌬2
⌬3
(24.10)
Equation (24.10) can be used in Equation (24.6). However, since the threedimensional geodetic network adjustment is performed in the local geodetic
coordinate system, the rotational elements of Equation (24.10) can be transformed about a single station in this system. Dropping the ⌬ symbol, the
rotation about the single station in the local geodetic system becomes
R ⫽ RT3 (0)R2T(90 ⫺ 0)R3(3)R2(2)R1(1)R2(90 ⫺ 0)R3(0)
(24.11)
where n is a rotation about the north axis of the local geodetic coordinate
system; e a rotation about the east axis of the local geodetic coordinate
system; u a rotation about the up axis of the local geodetic coordinate system;
R1, R2, and R3 are the rotation matrices defined in Section 18.7; and 0 and
0 are the geodetic coordinates of the rotational point. This point should be
picked near the center of the project area.
Again, since in nearly aligned coordinate systems the rotations are small,
the rotations above can be simplified to
(24.12)
R ⫽ nRn ⫹ eRe ⫹uRu ⫹ I
where
Ru ⫽
Re ⫽
Rn ⫽
0
⫺sin 0
cos 0 sin 0
0
0
cos 0
0
0
sin 0
0
cos 0
sin 0 sin 0
sin 0
0
⫺cos 0 cos 0
⫺cos 0 sin 0
cos 0 cos 0
0
⫺cos 0
⫺sin 0
0
⫺cos 0
0
⫺sin 0 sin 0
sin 0 cos 0
⫺sin 0 cos 0
0
I is a three-dimensional identity matrix; n, e, u are in radian units; and 0
and 0 are the geodetic coordinates of the rotational point.
486
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
Finally, the transformation going from the observed GPS vector to its local
geodetic equivalent between stations I and J is
⌬XIJ
⌬YIJ
⌬ZIJ
2
⌬XIJ
⫽ (1 ⫹ s)R ⌬YIJ
⌬ZIJ
(24.13)
1
where s represents the differential scale change between systems 1 and 2, R
is defined in Equation (24.12), [䡠]1 represents the GPS baseline vector components in the local coordinate system, and [䡠]2 represents the GPS observed
baseline vector components between stations I and J.
In Chapter 23 the three-dimensional geodetic network adjustment was developed in the local geodetic system. Thus, the addition of the GPS baseline
vectors into this adjustment requires that the coefficient matrix be rotated into
the same system. Recall from Chapter 17 that the coefficient matrix (A) for
each GPS baseline vectors consisted of three rows containing ⫺1, 0, and 1.
This matrix must be rotated into the local geodetic system. Thus, for the
baseline vector IJ, the new coefficient matrix values are derived as
⫺(1 ⫹ s)R RXG(i, i) RLG(i)⫺1
For station I:
(24.14)
⫺1
For station J:
(1 ⫹ s)R RXG(j, j) RLG(j)
where s is the change in scale between the two systems, rotation matrix R is
defined in Equation (24.12), rotation matrix RXG is defined in Equation (24.8),
rotation matrix RLG is defined in Equation (24.9), and (i, i) and (j, j) are
the geodetic coordinates from stations I and J, respectively.
When developing the matrices for a least squares adjustment, the unknown
parameters for scale and rotation should be set to zero. After the first iteration
these values will be modified and updated. At the end of the adjustment, these
parameters can be checked for statistical significance as described in Section
18.8.
The elements of the coefficient matrix for the unknown rotation angles and
scale are
XJ ⫺ XI rn1 re1 ru1
J ⫽ YJ ⫺ YI rn2 re2 ru2
ZJ ⫺ ZI rn3 re3 ru3
(24.15)
where X T ⫽ [⌬s ⌬n ⌬e ⌬u],
XJ ⫺ XI
rn ⫽ Rn YJ ⫺ YI ,
ZJ ⫺ ZI
XJ ⫺ XI
re ⫽ Re YJ ⫺ YI ,
ZJ ⫺ ZI
XJ ⫺ XI
ru ⫽ Ru YJ ⫺ YI
ZJ ⫺ ZI
24.4
COMBINING GPS BASELINE VECTORS WITH TRADITIONAL OBSERVATIONS
487
Example 24.2 The local datum coordinates for station B are given in Example 24.1 and the GPS-derived and transformed approximate NAD 83 coordinates for station E are (41⬚18⬘43.9622ⴖ N, 76⬚00⬘29.0384ⴖ W, 292.354 m).
Assuming initial values of zero for the scale and rotation parameters in Equation 24.13 and using point A as the rotational point, develop the coefficient
(J) and constants (K) matrices for the transformation parameter and the following GPS-observed baseline.
Baseline
⌬X (m)
⌬Y (m)
⌬Z (m)
EB
553.430
43.400
⫺62.949
SOLUTION From Example 24.1, the approximate NAD 83 geocentric and
geodetic coordinates for stations E and B are
Station
h (m)
E
B
A
41⬚18⬘43.9622ⴖ
41⬚18⬘40.4665ⴖ
41⬚18⬘26.0485ⴖ
⫺76⬚00⬘29.0385ⴖ
⫺76⬚00⬘05.5019ⴖ
⫺76⬚00⬘10.2486ⴖ
292.354
319.677
—
By Equation (24.8), RXG for station B is
⫺1016184.4632
RBXG ⫽
4076156.4570
4779908.9967
4655590.8060
0.1817
1160637.2558 ⫺0.7288
0.0000
0.6601
By Equation (24.8), RXG for station E is
⫺1015734.7290
REXG ⫽
4076334.1775
4779818.0865
4655634.2166
0.1816
1160083.8286 ⫺0.7288
0.0000
0.6602
By Equation (24.9), RLG for station B is
RBLG ⫽
6363584.8656
0.0000 0.0000
0.0000 4798083.4291 0.0000
0.0000
0.0000 1.0000
By Equation (24.9), RLG for station E is
488
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
RELG ⫽
6363558.6197
0.0000
0.0000
0.0000 0.0000
4797991.7099 0.0000
0.0000 1.0000
Since the initial approximations for the three differential rotations and scale
are initially 0, the R matrix in Equation (24.12) reduces to a 3 ⫻ 3 identity
matrix for the first iteration. In subsequent iterations, this matrix will change.
Combining the data from Table (23.3) with the GPS baseline vectors of
Example 24.1 yields a J matrix with 25 columns where the unknown parameters are [dXA dYA dZA dXB dYB dZB dXC 䡠 䡠 䡠 dZD dXE dYE dZE 䡠 䡠 䡠 s ne u]T. The J coefficient matrix for baseline EB is
J⫽
0 0 0 a11 a12 a13
0 0 0 a21 a22 a23
0 0 0 a31 a32 a33
0 䡠 䡠 䡠 0 a14
0 䡠 䡠 䡠 0 a24
0 䡠 䡠 䡠 0 a34
a15
a25
a35
a16
a26
a36
0 䡠 䡠 䡠 0 b11
0 䡠 䡠 䡠 0 b21
0 䡠 䡠 䡠 0 b31
b12
b22
b32
b13
b23
b33
b14
b24
b34
(24.16)
where, using Equation (24.14), the A-matrix coefficients are
⫺0.1597
A⫽
0.9703
0.1817
0.1596 ⫺0.9703 ⫺0.1816
0.6405 ⫺0.2419 ⫺0.7288 ⫺0.6406
0.2418
0.7288
0.7511
0.0000
0.6601 ⫺0.7511
0.0000 ⫺0.6601
and, using Equation (24.15), the B-matrix coefficients for the transformation
parameters are
B⫽
15.2308 ⫺17.2421
553.4271 ⫺72.9412
43.4106
405.6700 ⫺61.1002 ⫺376.7564
⫺62.9699 ⫺361.3992
91.7378 ⫺411.2673
With the aid of Equations (17.2) to (17.4), the geocentric coordinates for
stations E and B are
Station
X (m)
Y (m)
Z (m)
E
B
1,160,083.830
1,160,637.257
⫺4,655,634.217
⫺4,655,590.805
4,188,722.346
4,188,659.377
The values for the constants matrix (K) are
489
PROBLEMS
K⫽
553.430 ⫺ (1160637.257 ⫺ 1160083.830)
⫹0.003
43.400 ⫺ (⫺4655590.805 ⫹ 4655634.217) ⫽ ⫺0.012
⫺62.949 ⫺ (4188659.377 ⫺ 4188722.346)
0.020
Notice that the coefficient matrix for each GPS baseline vector component
is no longer a matrix of ⫺1, 0, and 1, but rather, contains noninteger values.
Also note that the coefficients for the differential changes in scale and rotations are populated for each baseline vector component. The reader may wish
to review the Mathcad worksheet on the CD that accompanies this book to
explore the complete set of matrix operations.
24.5
OTHER CONSIDERATIONS
Using procedures similar to those shown in Chapters 23 and Example 24.2,
a combined adjustment of the terrestrial and GPS baseline vectors can be
performed. If GPS-derived station coordinates are to be fixed during the adjustment, they must first be transformed in the local datum to ensure consistency with any local control stations. However, if GPS-derived points are the
only control in the adjustment, the entire adjustment can be performed using
the global datum that was used to reduce the GPS observations.
In both chapters the adjustments are performed in the local geodetic coordinate system. This system was chosen since defining standard deviations
for control stations in the (n,e,u) system is more intuitive to surveyors than
either the geodetic or geocentric coordinate systems. References in the bibliography contain procedures for combining GPS and terrestrial observations
using either the geocentric or geodetic systems.
PROBLEMS
24.1
Discuss what is meant by local datum.
24.2
Discuss what is meant by global datum.
24.3
How do datums differ?
24.4
Why is it important to perform a localization before staking out a
highway alignment?
24.5
What is meant by localization?
24.6
Using the Helmert transformation parameters derived in Example
24.1, derive the NAD 83 geocentric coordinates (in meters) for a
490
COMBINING GPS AND TERRESTRIAL OBSERVATIONS
point having ITRF 00 coordinates of (1160652.008, ⫺4655693.197,
4188423.986).
24.7
Repeat Problem 24.6 for a station having geocentric coordinates of
(1160398.043, ⫺4655803.184, 4188935.609).
24.8
Using the accompanying data, compute the Helmert transformation
parameters to take the coordinate values from WGS 84 to NAD 83.
NAD 83
WGS 84
Station
x (m)
y (m)
z (m)
x (m)
y (m)
z (m)
100
101
102
1,160,097.952
1,160,285.844
1,159,986.652
⫺4,634,583.300
⫺4,634,859.416
⫺4,634,623.501
4,188,086.049
4,188,233.622
4,188,153.783
1,160,098.356
1,160,286.248
1,159,987.056
⫺4,634,583.248
⫺4,634,859.364
⫺4,634,623.449
4,188,086.233
4,188,233.806
4,188,153.967
24.9
Why is it important to have control points dispersed about the perimeter of a project area during a stakeout survey?
24.10 Using the appropriate information from Table 23.3, Example 24.2,
and the accompanying baseline vector data, determine the nonzero
elements of the coefficient (J) and constant (K) matrices.
Baseline
⌬X (m)
⌬Y (m)
⌬Z (m)
ED
35.2573
⫺368.067
⫺347.063
24.11 Repeat Problem 24.10 for the following data. The approximate geodetic coordinates for station F are (41⬚18⬘39.7004ⴖ N, 76⬚59⬘58.9973ⴖ
W, 312.731 m).
Baseline
⌬X (m)
⌬Y (m)
⌬Z (m)
FB
⫺149.874
⫺25.079
22.222
24.12 Repeat Problem 24.10 for the following data. The approximate geodetic coordinates for station G are (41⬚18⬘12.8871ⴖ N, 76⬚00⬘11.2922ⴖ
W, 350.935 m).
Baseline
⌬X (m)
⌬Y (m)
⌬Z (m)
GD
⫺529.004
188.868
334.427
PROBLEMS
491
24.13 Using station A as the central point in a project and the baseline
vectors given in Problem 24.10, what are the first iteration coefficients
for the transformation parameters?
24.14 Repeat Problem 24.12 using the baseline from Problem 24.11.
24.15 Repeat Problem 24.12 using the baseline from Problem 24.12.
Programming Problem
24.16 Develop software that solves Problem 24.8.
CHAPTER 25
ANALYSIS OF ADJUSTMENTS
25.1
INTRODUCTION
An initial run of a least squares adjustment is not the end of the adjustment.
Rather, it is the beginning of the data analysis process. Throughout this book,
the mechanics of performing a least squares adjustment properly have been
discussed. Additionally, statistical methods have been introduced to analyze
the quality of observations. In this chapter we explore the procedures used in
analyzing the results of an adjustment beginning with the analysis of residuals
and reviewing data snooping as discussed in Chapter 21.
25.2 BASIC CONCEPTS, RESIDUALS, AND THE
NORMAL DISTRIBUTION
The normal distribution and statistical testing were introduced in the beginning chapters using simple data sets. These basic concepts also apply in the
analysis of data after an adjustment. When viewing Figure 25.1, the guiding
principles used in analyzing observations from normally distributed data are:
1.
2.
3.
4.
492
Data tend to be clustered around a single value.
Errors tend to be equally distributed about this value.
Errors are equally distributed in sign.
Most errors tend to be small in magnitude and errors large in magnitude
seldom occur.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
25.2
BASIC CONCEPTS, RESIDUALS, AND THE NORMAL DISTRIBUTION
493
Figure 25.1 Normal distribution.
At the conclusion of a least squares adjustment, the residuals in the observations should be scanned and analyzed. However, terms such as large in
magnitude must be defined statistically. This is accomplished by statistically
comparing the observation’s a priori standard deviation against its residual
using a t-distribution multiplier. For example, if a distance observation has a
residual of ⫹0.25 ft when all other residuals in the adjustment are below 0.10
ft, this observation may be viewed as large. However, if the observation had
an estimated standard deviation of only 0.5 ft, its residual is actually well
within its predicted range.
As an example, the results of the adjustment from Example 16.2 will be
analyzed. A quick review of the residuals listed for the distance observations
in Figure 25.2 show that the residual for distance QR is at least twice the size
of the other distance residuals and might be viewed as large. However, the a
priori standard deviation of this observation was 0.026 ft from an adjustment with 13 redundant observations. The 95% confidence interval for this
observational residual is computed as
v95 ⫽ t0.025,13S ⫽ 2.16(0.026) ⫽ 0.056 ft
From this it can be shown that the residual of ⫺0.038 ft for distance QR is
well within the 95% confidence interval of 0.056 ft.
When analyzing the angles, it can be seen QTR has a residual of 18.52ⴖ.
This residual is more than three times larger than the next largest residual of
5.23ⴖ. Also note that this angle opposes distance QR. Therefore it may explain
why distance QR had the largest distance residual. The a priori estimated
standard deviation for angle QTR is 4.0ⴖ. Using the same t-distribution
critical value for a 95% confidence interval yields
v95 ⫽ 2.16(4.0ⴖ) ⫽ 8.6ⴖ
494
ANALYSIS OF ADJUSTMENTS
*******************************
Adjusted Distance Observations
*******************************
Station
Station
Occupied
Sighted
Distance
V
S
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
Q
R
1,639.978
⫺0.0384
0.0159
R
S
1,320.019
0.0176
0.0154
S
T
1,579.138
0.0155
0.0158
T
Q
1,664.528
0.0039
0.0169
Q
S
2,105.953
⫺0.0087
0.0156
R
T
2,266.045
0.0104
0.0163
***************************
Adjusted Angle Observations
***************************
Station
Station
Station
Backsighted Occupied Foresighted
Angle
V
S
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
R
Q
S
38⬚4852.8ⴖ
2.08ⴖ 1.75
S
Q
T
47⬚4613.1ⴖ
0.70ⴖ 1.95
T
Q
R
273⬚2454.1ⴖ ⫺2.37ⴖ 2.40
Q
R
S
269⬚5734.0ⴖ
0.64ⴖ 2.26
R
S
T
257⬚3257.3ⴖ
0.50ⴖ 2.50
S
T
Q
279⬚0434.5ⴖ
3.32ⴖ 2.33
S
R
T
42⬚5252.3ⴖ
1.34ⴖ 1.82
S
R
Q
90⬚0226.0ⴖ ⫺0.74ⴖ 2.26
Q
S
R
51⬚0841.3ⴖ ⫺3.73ⴖ 1.98
T
S
Q
51⬚1821.4ⴖ
5.23ⴖ 2.04
Q
T
R
46⬚1520.5ⴖ
18.52ⴖ 1.82
R
T
S
34⬚4005.0ⴖ ⫺0.74ⴖ 1.72
Figure 25.2 Adjusted distances and angles from Example 16.2.
The actual residual for angle QTR is well outside the range of 8.6ⴖ. In
this case, the residual for angle QTR definitely fits the definition of large and
it is a candidate for reobservation or removal from the adjustment.
Using the procedures outlined in Chapter 21, data snooping confirms that
angle QTR is a detectable blunder. This angular observation can be removed
from the data set and the adjustment rerun. The results of this readjustment
are shown in Figure 25.3. Note that the residual for distance QR was reduced
by about half of its previous value. This demonstrates the interrelationship
between the distance QR and angle QTR. Also note that all the residuals are
within a single standard deviation of their a priori estimated values. In fact,
all the distance residuals are less than a third of their a priori estimated errors,
and all the angle residuals are less than half their a priori estimates.
There are 18 observations in this adjustment. Thus, about half of the residuals (nine) should be positive and the others negative. In this adjustment
25.2
BASIC CONCEPTS, RESIDUALS, AND THE NORMAL DISTRIBUTION
495
Adjusted stations
Error ellipse confidence level at 0.950
Station
X
Y
Sx
Sy
Su
Sv
t
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
R
1,003.06 2,640.01 0.000 0.006 0.017 0.000
0.11⬚
S
2,323.06 2,638.47 0.005 0.007 0.019 0.014 156.28⬚
T
2,661.74 1,096.09 0.006 0.007 0.021 0.015
26.18⬚
Adjusted Distance Observations
Station
Station
Occupied Sighted Distance
V
S
Std.Res. Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
Q
R
1,640.008 ⫺0.0081 0.0060
⫺0.409
0.576
R
S
1,320.006
0.0054 0.0055
0.295
0.579
S
T
1,579.133
0.0099 0.0056
0.510
0.597
T
Q
1,664.514 ⫺0.0097 0.0060
⫺0.495
0.569
Q
S
2,105.966
0.0039 0.0056
0.162
0.702
R
T
2,266.034 ⫺0.0014 0.0058
⫺0.057
0.700
Adjusted Angle Observations
Station
Station
Station
Backsighted Occupied Foresighted
Angle
V
S Std.Res. Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
R
Q
S
38⬚4850.2ⴖ ⫺0.45ⴖ 0.64 ⫺0.13 0.795
S
Q
T
47⬚4611.7ⴖ ⫺0.73ⴖ 0.69 ⫺0.21 0.757
T
Q
R
273⬚2458.1ⴖ 1.58ⴖ 0.89
0.44 0.672
Q
R
S
269⬚5734.7ⴖ 1.31ⴖ 0.80
0.32 0.767
R
S
T
257⬚3256.9ⴖ 0.11ⴖ 0.88
0.03 0.716
S
T
Q
279⬚0430.3ⴖ ⫺0.91ⴖ 0.87 ⫺0.24 0.700
S
R
T
42⬚5252.6ⴖ 1.58ⴖ 0.64
0.41 0.821
S
R
Q
90⬚0225.3ⴖ ⫺1.41ⴖ 0.80 ⫺0.36 0.746
Q
S
R
51⬚0844.5ⴖ ⫺0.53ⴖ 0.73 ⫺0.14 0.767
T
S
Q
51⬚1818.6ⴖ 2.43ⴖ 0.74
0.71 0.722
R
T
S
34⬚4004.3ⴖ ⫺1.37ⴖ 0.61 ⫺0.38 0.814
Adjusted Azimuth Observations
Station
Station
Occupied
Sighted
Azimuth
V
Sⴖ
Std.Res.
Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
Q
R
0⬚0624.5ⴖ
⫺0.00ⴖ
0.00ⴖ
0.0
0.000
Adjustment Statistics
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
Iterations ⫽ 2
Redundancies ⫽ 12
Reference Variance ⫽ 0.1243
Reference So ⫽ 0.35
Failed to pass X2 test at 95.0% significance level!
X2 lower value ⫽ 4.40
X2 upper value ⫽ 23.34
Figure 25.3 Readjustment of the data in Example 16.2 after removing angle QTR.
496
ANALYSIS OF ADJUSTMENTS
there are exactly nine positive residuals and nine negative residuals. This
follows the second guideline for normally distributed data. In this example,
the residuals do not show any skewness, as discussed in Section 2.4. However,
it should be remembered that since we are working with a sample of data,
some variation from theory is acceptable.
Example 25.1 A GPS baseline vector determined by the rapid-static method
(5 mm ⫹ 1 ppm) has a length of about 10.5 km. The instrument setups were
estimated to be within 0.003 mm of the true station location. Its combined
XYZ residual is 0.021 m. At a 95% level of confidence, should this residual
be considered too large if the number or redundancies in the adjustment is
20?
SOLUTION The a priori estimated error in this baseline using the rapidstatic method is
⫽
冪
32 ⫹ 32 ⫹ 52 ⫹
冉
冊
10,500,000
1,000,000
2
⫽ 12.4 mm
t0.025,20 ⫽ 2.09
The acceptable 95% range is R ⫽ 2.09(12.4) ⫽ 25.9 mm. The residual
was 21 mm. This is inside of the 95% confidence interval and thus is an
acceptable size for a combined residual vector with this baseline and surveying method.
25.3
GOODNESS-OF-FIT TEST
The results of the adjustment in Figure 25.3 appear to follow the normal
distribution and the size of the residuals indicates that the data appear to be
consistent. However, the adjustment failed to pass the goodness-of-fit (2) test.
This demonstrates a weakness in the goodness-of-fit test. That is, passing or
failing the test is not a good indicator of the quality of the data or the presence
of blunders. Nor does the test reveal the exact problem in data when the test
fails. The 2 test should be viewed as a warning flag for an adjustment that
requires further analysis, not as an indicator of bad data.
As discussed in Chapters 5 and 16, the 2 test compares the a prior reference variance, set equal to 1, against the reference variance computed from
the data. An analysis of Equation (12.15) shows that the test will fail if the
residuals are too large or too small compared to the weights of the observations. For example, an observation with a small residual should have a high
weight and a large residual should have a low weight. If the residuals tend
to be smaller than their a priori standard deviations, the resulting reference
variance will probably be less than 1. This is an example of an incorrect
stochastic model. If the residuals tend to be larger than their a priori standard
25.3
GOODNESS-OF-FIT TEST
497
deviations, the reference variance will be greater than 1 and the 2 test will
fail in the upper bounds of the distribution.
The original data set of Example 16.2 failed the goodness-of-fit test. In the
original adjustment, it failed because the computed reference variance was
greater than 1. This was probably caused by the large residual for angle QTR.
After angle QTR was removed from the data, the 2 test failed because the
reference variance was statistically less than 1. Since the residuals appear to
be much smaller than their a priori standard deviations, an incorrect stochastic
model is to blame. In this case, the distance residuals were about three times
less than the estimated standard deviations and the angles were one-half of
their estimated standard deviations. Thus, to pass the 2 test with this adjustment, the stochastic model needs to be modified.
In developing the stochastic model for this problem, it was originally believed that the field crew had use a total station with a DIN 18723 accuracy
of 4ⴖ and an EDM accuracy of 5 mm ⫹ 5 ppm. It was later discovered that
the crew used an instrument having a DIN 18723 accuracy of 2ⴖ and an
EDM accuracy of 2 mm ⫹ 2 ppm. Estimated standard deviations using this
new information resulted in the adjustment shown in Figure 25.4. This adjustment passed the goodness-of-fit test. However, this did not change the
adjustment results significantly. That is, the adjusted coordinate values and
adjusted observations are nearly the same in the adjustments shown in Figures
25.3 and 25.4. What did change in Figure 25.4 were the statistical results.
Since the statistical results of the adjustment are of little value to the surveyor
once analyzed, the adjustment shown in Figure 25.3 is sufficient for most
applications.
Practitioners often view the goodness-of-fit test as an indicator of a possible
problem. If the goodness-of-fit test failed because the computed reference
variance was too small, the test result is often ignored, since correcting it
does not change the adjusted coordinates significantly. In this case, some
practitioners will say that the adjustment failed the goodness-of-fit test on the
‘‘good’’ side. That is, the reference variance was too small because the residuals are too small for their given weights. However, if the reference variance
computed is greater than 1, the results should be analyzed for possible blunders in the observations. After the identifiable blunders have been eliminated
from the data set, the stochastic model might still exhibit a problem. In this
case, if the computed reference variance is greater than 1, the estimated standard deviations for the observations should be increased. In the preceding
example the computed reference variance is less than 1. To correct this, the
estimated standard deviations for the observations were decreased.
This adjustment brings about another important discussion point. Had all
the a priori standard deviations been reduced by a single scale factor, the
results of the adjustment would be the same.1 That is, if the a priori standard
1
Readers are encouraged to rerun the data in Example 16.2 with the a priori standard deviations
reduced by a factor of –12 to compare the adjustment results.
498
ANALYSIS OF ADJUSTMENTS
Adjusted stations
Error ellipse confidence level at 0.950
Station
X
Y
Sx
Sy
Su
Sv
t
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
R
1,003.06 2,640.00 0.000 0.009 0.024 0.000
0.11⬚
S
2,323.06 2,638.47 0.008 0.009 0.028 0.021 151.27⬚
T
2,661.74 1,096.09 0.008 0.011 0.031 0.022
26.14⬚
Adjusted Distance Observations
Station
Station
Occupied Sighted Distance
V
S
Std.Res. Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
Q
R
1,640.008 ⫺0.0083 0.0086
⫺0.830
0.580
R
S
1,320.006
0.0053 0.0081
0.527
0.603
S
T
1,579.133
0.0100 0.0081
0.971
0.615
T
Q
1,664.515 ⫺0.0089 0.0087
⫺0.895
0.568
Q
S
2,105.966
0.0039 0.0079
0.341
0.673
R
T
2,266.034 ⫺0.0011 0.0082
⫺0.097
0.658
Adjusted Angle Observations
Station
Station
Station
Backsighted Occupied Foresighted
Angle
V
S Std.Res. Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
R
Q
S
38⬚4850.2ⴖ ⫺0.46ⴖ 0.95 ⫺0.24 0.808
S
Q
T
47⬚4611.7ⴖ ⫺0.73ⴖ 1.02 ⫺0.38 0.776
T
Q
R
273⬚2458.1ⴖ 1.59ⴖ 1.29
0.87 0.667
Q
R
S
269⬚5734.7ⴖ 1.27ⴖ 1.14
0.64 0.751
R
S
T
257⬚3256.8ⴖ 0.03ⴖ 1.27
0.02 0.696
S
T
Q
279⬚0430.4ⴖ ⫺0.79ⴖ 1.25 ⫺0.43 0.684
S
R
T
42⬚5252.6ⴖ 1.55ⴖ 0.94
0.77 0.820
S
R
Q
90⬚0225.3ⴖ ⫺1.37ⴖ 1.14 ⫺0.69 0.751
Q
S
R
51⬚0844.4ⴖ ⫺0.57ⴖ 1.06 ⫺0.29 0.772
T
S
Q
51⬚1818.7ⴖ 2.53ⴖ 1.09
1.34 0.751
R
T
S
34⬚4004.3ⴖ ⫺1.42ⴖ 0.90 ⫺0.72 0.827
Adjusted Azimuth Observations
Station
Station
Occupied
Sighted
Azimuth
V
Sⴖ
Std.Res.
Red.#
⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽⫽
Q
R
0⬚0624.5ⴖ
⫺0.00ⴖ
0.00ⴖ
0.0
0.000
Adjustment Statistics
*********************
Iterations ⫽ 2
Redundancies ⫽ 12
Reference Variance ⫽ 0.4572
Reference So ⫽ 0.68
Passed X2 test at 95.0% significance level!
X2 lower value ⫽ 4.40
X2 upper value ⫽ 23.34
Figure 25.4 Readjusted data from Example 16.2 with a different stochastic model.
25.4
COMPARISON OF RESIDUAL PLOTS
499
deviations for all the observations had been cut in half, this would have resulted in increasing all the weights by a factor of 4. Since weights are relative,
performing a scalar reduction in the standard deviations does not change the
adjustment.
This example demonstrates the importance of selecting appropriate a priori
standard deviations for the observations. Estimated standard deviations for
observations cannot be selected from the recesses of one’s mind. Doing this
not only affects how the errors in the observations are distributed but also
how the results of the adjustment are analyzed. For traditional surveys it is
always best to select a stochastic model that reflects the estimated accuracies
in the observations using procedures discussed in Chapters 7 and 9. Good
estimates of setup errors will typically result in a sound stochastic model.
However, the network adjustment of GPS baseline vectors derives its stochastic model from the least squares reduction of each baseline. As discussed
in Chapter 17, the stochastic model is part of the printout from the baseline
reduction. Assuming proper field procedures, if the GPS network adjustment
fails to pass the 2 test, there is little one can do to modify the model. In
fact, if field procedures are consistent, scaling the entire stochastic model
might be the only option. Although this procedure may result in a ‘‘passed’’
2 test, it will not change the coordinates in the solution since weights are
relative. In fact, it will only change the a posteriori statistics. Thus, scaling
the entire stochastic model by a single factor is important only if the statistics
are important. Since this is seldom true, there is little value in scaling the
stochastic model.
Better adjustments could be obtained if the commercial GPS software employed the multipoint solution technique. This solution takes the individual
pseudoranges from the satellites and computes baseline vectors and the unknown coordinates of the network stations simultaneously. Current software
offerings compute the baseline vectors from the pseudoranges and then coordinates of the network stations from these computed vectors. This is a twostep process. The multipoint solution is similar to the three-dimensional
geodetic network and the photogrammetric block-bundle adjustment in that
all unknowns are computed in one adjustment. However, it is important to
remove any large blunders before attempting a multipoint solution. Thus, the
two-step solution process must be maintained to ferret out problems in the
data. Since the main advantage of the multipoint solution lies in the development of the stochastic model, only a few examples of this solution technique exist.
25.4
COMPARISON OF RESIDUAL PLOTS
Massive quantities of data can be collected and reduced during the reduction
of GPS carrier-phase observations. Software manufacturers typically plot
pseudorange residuals against a time line. Examples of this are shown in
500
ANALYSIS OF ADJUSTMENTS
Figure 25.5, where plot (a) is for satellite vehicle 24 (SV 24) and (b) is for
satellite vehicle 28 (SV 28). Notice that the residuals in Figure 25.5(b) have
a slight downward slope at the beginning of the session. In fact, except for
the first 3 minutes, the residuals are fairly consistent even though most are
negative in sign. During the first 3 minutes of the session, satellite 28 was
just clearing the set mask angle and suffered some loss of lock problems.
This skewed the data.
The first 3 minutes of data from this satellite could be removed and the
baseline reprocessed to correct this problem. However, this might not be practical given the shortness of the overall observation session. Since much field
time is lost to travel, setup, and teardown, it is always wise to collect more
data than are needed to resolve the position of the receiver. This is similar to
observing distances and angles more than once.
Note that the residuals in Figure 25.5(a) are mostly positive. This could
have several causes, including loss-of-lock problems. However, these types
(a)
(b)
Figure 25.5 Pseudorange residual plots from satellites 24 (a) and 28 (b) versus time.
25.5
USE OF STATISTICAL BLUNDER DETECTION
501
of plots are typically seen when the broadcast ephemeris is used in the reduction. The broadcast ephemeris is a near-future prediction of the location
of the satellite. For various reasons, satellites often stray from their predicted
paths. When this happens, the coordinates for the satellite’s position at the
time of broadcast are in error. This problem results in plots like Figure
25.5(a), where most of the residuals typically have one sign indicating a
skewed data set. This problem can be corrected by downloading and processing with a precise ephemeris. In general, the best solutions can always
be obtained with a precise ephemeris. Since the ultrarapid ephemeris is available within a few hours of data collection, it is wise to wait for this ephemeris
before processing the data.
Figure 25.5 also depicts a problem with real-time kinematic GPS (RTKGPS) surveys. That is, by its very nature, the broadcast ephemeris must be
used in processing the data. Thus, the option of downloading a precise ephemeris is not available. RTK-GPS has its place in stakeout and mapping surveys.
However, it should never be used to establish anything other than very loworder control. The RTK-GPS survey is the equivalent of a radial total station
survey, leaving only minimal checks in the quality of the data. Since all GPS
surveys can be seriously affected by solar activity, the RTK-GPS procedure
leaves the surveyor open to poor results without any indication at the receiver.
25.5
USE OF STATISTICAL BLUNDER DETECTION
When more than one blunder exists in a given data set, the statistical blunder
detection methods presented in Chapter 21 help isolate the offending observations. These methods along with graphical techniques, were presented thoroughly in that chapter.
The problem of having multiple blunders usually exists in large control
surveys. When adjusting large data sets, it is wise to break the data into
smaller subsets that can be checked for blunders before attempting the larger
adjustment. By breaking a larger adjustment into smaller subsets of data, data
problems can often be isolated, corrected, or removed from the observations.
GPS provides a similar example. When performing a large GPS adjustment, it is often wise initially to download and process GPS data by session.
Since a GPS campaign typically involves several observational sessions, trying to combine all sessions into one large adjustment generally takes more
time than individually processing the data by session. Processing GPS data
by session simplifies the process of isolating poor and trivial baselines. After
processing each session in a GPS survey satisfactorily, a combined network
adjustment can be attempted and analyzed for blunders. It may be important
at this point to reanalyze and possibly reprocess baselines to obtain the best
solution. Again, statistical blunder detection will help analyze possible blunders in baselines. If the observation sessions are long enough, it is possible
to eliminate data from the session to remove possible problems caused by
502
ANALYSIS OF ADJUSTMENTS
obstructions and loss-of-lock problems. This again indicates the need to collect more data than are minimally required to obtain a solution. A few extra
minutes used to observe additional data can often result in improved solutions
without returning to the field for reobservation.
PROBLEMS
25.1
Following an adjustment, how is the term large defined in relation to
residuals?
25.2
What are the basic concepts of a normal distribution, and how are
these concepts used to analyze the results of a least squares adjustment?
25.3
An adjustment has 20 observations. Only nine of these residuals are
negative in sign. Should this cause concern during the analysis?
25.4
Discuss why the goodness-of-fit test is not always a reliable indicator
of a blunder in data.
25.5
What does it mean to ‘‘fail on the good side’’ of the 2 test?
25.6
Discuss the importance of a proper stochastic model.
25.7
Discuss why the scaling of the stochastic model does not affect the
adjusted parameters.
25.8
What is the possible reason for the residual plot in Figure 25.5a?
25.9
Discuss the importance of collecting more than the minimum amount
of data in a GPS survey.
25.10 Why should the RTK-GPS method not be used to establish high-order
control stations?
25.11 Why should GPS baseline vectors be processed by session rather than
by job?
25.12 A GPS baseline vector determined by the rapid-static method (5 mm
⫹ 1 ppm) has a length of about 5.5 km. The instrument setups were
estimated to be within 0.003 m of the true station location. Its
combined XYZ residual is 2.5 cm. At a 95% confidence level, should
this residual be considered too large if the number redundancies in
the adjustment is 15?
In Problems 25.13 through 25.20, analyze the results of the least squares
adjustment, indicating if there are any areas of concern in the data or stochastic model. When possible, isolate and remove questionable data and stochastic models. After the changes, rerun the adjustment and compare the
results. (Note: Not all of the problems have questionable data.)
PROBLEMS
25.13 Problem 12.1
25.14 Problem 12.13
25.15 Problem 13.15
25.16 Problem 15.9
25.17 Problem 16.7
25.18 Problem 16.9
25.19 Problem 16.11
25.20 Problem 17.8
503
CHAPTER 26
COMPUTER OPTIMIZATION
26.1
INTRODUCTION
Large amounts of computer time and storage requirements can be used when
performing least squares adjustments. This is due to the fact that as the problems become more complex, the matrices become larger, and the storage
requirements and time consumed in doing numerical operations both grow
rapidly. As an example, in analyzing the storage requirements of a 25-station
horizontal least squares adjustment that has 50 distance and 50 angle observations, the coefficient matrix would have dimensions of 100 rows and 50
columns. If this adjustment were done in double precision,1 it would require
40,000 bytes of storage for the coefficient matrix alone. The weight matrix
would require an additional 80,000 bytes of storage. Also, at least two additional intermediate matrices2 must be formed in computing the solution.
From this example it is easy to see that large quantities of computer time and
computer memory can be required in least squares adjustments. Thus, when
writing least squares software, it is desirable to take advantage of some storage and computing optimization techniques. In this chapter some of these
techniques are described.
26.2
STORAGE OPTIMIZATION
Many matrices used in a surveying adjustment are large but sparse. Using the
example above, a single row of the coefficient matrix for a distance obser1
The storage requirements of a double-precision number is 8 bytes.
The intermediate matrices developed are AT and ATW or J T and J TW, depending on whether the
adjustment is linear or nonlinear.
2
504
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
26.2
STORAGE OPTIMIZATION
505
vation would require a 50-element row for its four nonzero elements. In fact,
the entire coefficient matrix is very sparsely populated by nonzero values.
Similarly, the normal matrix is always symmetric, and thus nearly half its
storage is used by duplicate entries. It is relatively easy to take advantage of
these conditions to reduce the storage requirements.
For the normal matrix, only the upper or lower triangular portion of the
matrix need be saved. In this storage scheme, the two-dimensional matrix is
saved as a vector. An example of a 4 ⫻ 4 normal matrix is shown in Figure
26.1. Its upper and lower triangular portions are shown separated and their
elements numbered for reference. The vector on the right of Figure 26.1
shows the storage scheme. This scheme eliminates the need to save the duplicate entries of the normal matrix but requires some form of relational
mapping between the original matrix indices of row i and column j and the
vector’s index of row i. For the upper triangular portion of the matrix, it can
be shown that the vector index for any (i,j) element is computed as
Index(i,j) ⫽ –12 [ j ⫻ ( j ⫺ 1)] ⫹ i
(26.1)
Using the function given in Equation (26.1), the vector index for element
(2,3) of the normal matrix would be computed as
Index(2,3) ⫽ –12 [3 ⫻ (3 ⫺ 1)] ⫹ 2 ⫽ 5
An equivalent mapping formula for the lower triangular portion of the matrix
is given as
Index(i,j) ⫽ –12 [i ⫻ (i ⫺ 1)] ⫹ j
(26.2)
Equation (26.1) or (26.2) can be used to compute the storage location for
each element in the upper or lower part of the normal matrix, respectively.
Of course, it requires additional computational time to map the location of
each element every time it is used. One method of minimizing computing
Figure 26.1 Normal matrix.
506
COMPUTER OPTIMIZATION
time is to utilize a mapping table. The mapping table is an integer vector that
identifies the storage location for only the initial element of a row or column.
To reduce total computational time, the indices are stored as one number less
than their actual value. The example matrix stored in Figure 26.1 would have
a relational mapping table of
VI T ⫽ [0 1 3 6 10]
Although the storage location of the first element is 1, it can be seen above
that it is identified as 0. Similarly, every other mapping index is one number
less than its actual location in the vector, as shown in Figure 26.1. The mapping vector is found using the pseudocode shown in Table 26.1.
By using this type of mapping table, the storage location for the first element of any upper triangular column is computed as
Index(i,j) ⫽ VI( j) ⫹ i
(26.3)
Using the mapping table, the storage location for the (2,3) element is
Index(2,3) ⫽ VI(3) ⫹ 2 ⫽ 3 ⫹ 2 ⫽ 5
The equivalent formula for the lower triangular portion is
Index(i,j) ⫽ VI(i) ⫹ j
(26.4)
Obviously, this method of indexing an element’s position requires additional storage in a mapping table, but the additional storage is offset by the
decreased computational time that is created by the operations equation (26.1)
or (26.2). Since the mapping table replaces the more time-consuming multiplication and division operations of Equations (26.1) and (26.2), it is prudent
to use this indexing method whenever memory is available. Although every
compiler and machine are different, a set of times to compare indexing methods are shown in Table 26.2. The best times were achieved by minimizing
both the number of operations and calls to the mapping function. Thus, in
method C the mapping table was created to be a global variable in the program, and the table was accessed directly as needed. For comparison, the time
to access the matrix elements directly using the loop shown in method D was
1.204 seconds. Thus, method C uses only slightly more time than method D
TABLE 26.1 Creation of a Mapping Table
VI(1) ⫽ 0
For i going from 2 to the number of unknowns: VI(i) ⫽ VI(i ⫺ 1) ⫹ i ⫺ 1
26.3
DIRECT FORMATION OF THE NORMAL EQUATIONS
507
TABLE 26.2 Comparison of Indexing Methods
Method
Time
(sec)
Extra Storage
A
B
C
D
4.932
3.937
1.638
1.204
0
2 bytes per element
2 bytes per element
Full matrix
Function A: Index ⫽ j( j ⫺ 1) / 2 ⫹ i
Function B: Index ⫽ VI( j) ⫹ i;
Method A: Using function A
i from 1 to 1000
j from i to 1000
k ⫽ Index(i,j)
Method B: Using function B
i from 1 to 1000
j from i to 1000
k ⫽ Index(i,j)
Method C: Direct access to mapping table
i from 1 to 1000
j from i to 1000
k ⫽ VI( j) ⫹ I
Method D: Direct matrix element access.
i from 1 to 1000
j from 1 to 1000
k ⫽ A[i,j]
but helps avoid the additional memory that is required by the full matrix
required for method D. Meanwhile, method A is slower than method C by a
factor of 3. Since the matrix elements must be accessed repeatedly during a
least squares solution, this difference can be a significant addition to the total
solution time.
The additional overhead cost of having the mapping table is minimized by
declaring it as an integer array so that each element requires only 2 bytes of
storage. In the earlier example of 25 stations, this results in 100 bytes of
additional storage for the mapping table. Speed versus storage is a decision
that every programmer will face continually.
26.3
DIRECT FORMATION OF THE NORMAL EQUATIONS
The basic weighted observation equation form is WAX ⫽ WL ⫹ WV. The
least squares solution for this equation is X ⫽ (ATWA)⫺1(ATWL) ⫽ N⫺1C,
where ATWA is the normal equations matrix N, and ATWL is the constants
matrix C. If the weight matrix is diagonal, formation of the A, W, L, and ATW
matrices is unnecessary when building the normal equations, and the storage
requirements for them can be eliminated. This is accomplished by forming
the normal matrix directly from the observations. The tabular method in Section 11.8 showed the feasibility of this method. Notice in Table 11.2 that the
contribution of each observation to the normal matrix is computed individually and, subsequently, added. This shows that there is no need to form the
coefficient, constants, weight, or any intermediate matrices when deriving the
normal matrix. Conceptually, this is developed as follows:
508
COMPUTER OPTIMIZATION
Step 1: Zero the normal and constants matrix.
Step 2: Zero a single row of the coefficient matrix.
Step 3: Based on the values of a single row of the coefficient matrix, add the
proper value to the appropriate normal and constant elements.
Step 4: Repeat steps 3 and 4 for all observations.
This procedure works in all situations that involve a diagonal weight matrix. Procedural modifications can be developed for weight matrices with limited correlation between the observations. Computer algorithms in BASIC,
FORTRAN, C, and Pascal for this method are shown in Table 26.3.
26.4
CHOLESKY DECOMPOSITION
Having formed only a triangular portion of the normal matrix, a Cholesky
decomposition can be used to greatly reduce the time needed to find the
solution. This procedure takes advantage of the fact that the normal matrix is
always a positive definite3 matrix. Due to this property, it can be expressed
as the product of a lower triangular matrix and its transpose; that is,
N ⫽ LU ⫽ LLT ⫽ U TU
where L is a lower triangular matrix of the form
l11 0 0 䡠 䡠 䡠
l21 l22 0 䡠 䡠 䡠
L ⫽ l31 l32 l33
⯗
⯗
⯗
ln1 ln2 ln3
0
0
0
䡠䡠䡠 0
䡠 䡠 䡠 lnn
When the normal matrix is stored as a lower triangular matrix, it can be
factored using the following procedure. For i ⫽ 1, 2, . . . , number of unknowns, compute
冘l )
i⫺1
lii ⫽ (lii ⫺
2 1/2
ik
(26.5)
k⫽1
For j ⫽ i ⫹ 1, i ⫹ 2, . . . , number of unknowns, compute
i⫺1
lji ⫺ 兺k⫽1
lijljk
lji ⫽
lii
The procedures above are shown in code form in Table 26.4.
3
A positive definite square matrix, A, has the property that X TAX ⬎ 0 for all nonzero vectors, X.
TABLE 26.3 Algorithms for Building the Normal Equations Directly from Their Observations
BASIC Language:
For i ⫽ 1 to unknown
ix ⫽ VI(i): ixi ⫽ ix ⫹ i
N(ixi) ⫽ N(ixi) ⫹ A(i)2 * W(i)
C(i) ⫽ C(i) ⫹ A(i)*W(i)*L(i)
For j ⫽ i⫹1 to unknown
N(ix⫹j) ⫽ N(ix⫹j) ⫹ A(i)*W(i)*A( j)
Next j
Next I
FORTRAN Language:
Do 100 i ⫽ 1, unknown
ix ⫽ VI(i)
ixi ⫽ ix ⫹ i
N(ixi) ⫽ N(ixi) ⫹ A(i)**2 * W(i)
C(i) ⫽ C(i) ⫹ A(i)*W(i)*L(i)
Do 100 j ⫽ i⫹1, unknown
N(ix⫹j) ⫽ N(ix⫹j) ⫹ A(i)*W(i)*A( j)
100 Continue
C Language:
for (i⫽1; i⬍⫽unknown; i⫹⫹) {
ix ⫽ VI[i]; ixi ⫽ ix ⫹ i;
n[ixi] ⫽ n[ixi] ⫹ a[i]*a[i] * w[i];
c[i] ⫽ c[i] ⫹ a[i]*w[i]*l[i];
for ( j⫽i⫹1; j⬍⫽unknown; j⫹⫹)
n[ix⫹j] ⫽ n[ix⫹j] ⫹ a[i]*w[i]*a[j];
} / / for i
Pascal Language:
For i :⫽ 1 to unknown do begin
ix :⫽ VI[i]; ixi :⫽ ix ⫹ i;
N[ixi] :⫽ N[ixi] ⫹ Sqr(A[i]) * W[i];
C[i] :⫽ C[i] ⫹ A[i]*W[i]*L[i];
For j :⫽ i⫹1 to unknown do
N[ix⫹j] :⫽ N[ix⫹j] ⫹ A[i]*W[i]*A[j]
End; {for i}
509
510
TABLE 26.4 Computer Algorithms for Computing Cholesky Factors of a Normal Matrix
BASIC Language:
FOR i ⫽ 1 TO unknown
ix ⫽ VI(i): ixi ⫽ ix ⫹ i: s ⫽ 0#
FOR k ⫽ 1 TO i ⫺ 1
s ⫽ s ⫹ N(ix ⫹ k) ˆ 2: NEXT k
N(ixi) ⫽ SQR(N(ixi) ⫺ s)
FOR j ⫽ i ⫹ 1 TO unknown
s ⫽ 0#: jx ⫽ VI( j)
FOR k ⫽ 1 TO i ⫺ 1
s ⫽ s ⫹ N( jx ⫹ k) * N(ix ⫹ k): NEXT k
N( jx ⫹ i) ⫽ (N( jx ⫹ i) ⫺ s) / N(ixi)
NEXT j
NEXT i
FORTRAN Language:
Do 30 i ⫽ 1,Unknown
ix ⫽ VI(i)
i1 ⫽ i ⫺ 1
S ⫽ 0.0
Do 10 k⫽1, i1
10
s ⫽ s ⫹ N(ix⫹k)**2
N(ix⫹i) ⫽ Sqrt(N(ix⫹i)⫺s)
Do 30 j ⫽ i⫹1, Unknown
s ⫽ 0.0
Do 20 k ⫽ 1, i1
20
s ⫽ s ⫹ N(VI[j]⫹k) * N(ix⫹k)
N(VI[j]⫹i) ⫽ (N(VI[j]⫹i) ⫺ s) / N(ix⫹i)
30 Continue
C Language:
for (i⫽1; i⬍⫽unknown; i⫹⫹) {
ix ⫽ VI[i]; ixi ⫽ ix⫹i; s ⫽ 0.0;
for (k⫽1; k⬍i; k⫹⫹)
s ⫽ s ⫹ n[ix⫹k]*n[ix⫹k];
n[ixi] ⫽ sqrt(n[ixi] ⫺ s);
for ( j⫽i⫹1; j⬍⫽unknown; j⫹⫹) {
s ⫽ 0; jx ⫽ VI[j];
for (k⫽1; k⬍i; k⫹⫹)
s ⫽ s ⫹ n[jx⫹k] * n[ix⫹k];
n[jx⫹i] ⫽ (n[jx⫹i] ⫺ s) / n[ixi];
} / / for j
} / / for I
Pascal Language:
For i :⫽ 1 to unknown do Begin
ix :⫽ VI[i]; ixi :⫽ ix⫹i; S :⫽ 0;
For k :⫽ 1 to Pred(i) do
S :⫽ S ⫹ Sqr(N[ix⫹k]);
N[ixi] :⫽ Sqrt(N[ixi] ⫺ S);
For j :⫽ Succ(i) to unknown do Begin
S :⫽ 0; jx :⫽ VI[j];
For k :⫽ 1 to Pred(i) do
S :⫽ S ⫹ N[jx⫹k]*N[ix⫹k];
N[jx⫹i] :⫽ (N[jx⫹i] ⫺ S) / N[ixi]
End; {for j}
End; {for i}
26.5
26.5
511
FORWARD AND BACK SOLUTIONS
FORWARD AND BACK SOLUTIONS
Being able to factor the normal matrix into triangular matrices has the advantage that the matrix solution can be obtained without the use of an inverse.
The equivalent triangular matrices representing the normal equations are
(ATWA)X ⫽ NX ⫽ LUX ⫽ LLTX
l11 0 0
l21 l22 0
⫽ l31 l32 l33
⯗
⯗
⯗
ln1 ln2 ln3
c1
c2
⫽ c3 ⫽ C
0
0
0
䡠䡠䡠 0
䡠 䡠 䡠 lnn
䡠䡠䡠
䡠䡠䡠
l11 l21
0 l22
0 0
⯗
⯗
0
0
l31
l32
l33
0
䡠
䡠
䡠
䡠
䡠
䡠
䡠
䡠
0
䡠
䡠
䡠
䡠
ln1
ln2
ln3
⯗
lnn
x1
x2
x3
⯗
xn
(26.6)
⯗
cn
Equation (26.6) can be rewritten as LY ⫽ C, where LTX ⫽ Y. From this,
the solution for Y can be found by taking advantage of the triangular form of
L. This is known as a forward substitution. Steps involved in forward substitution are as follows:
Step 1: Solve for y1 as y1 ⫽ c1 /l11.
Step 2: Substitute this value into row 2, and compute y2 as y2 ⫽ (c2 ⫺ l21y1)/
l22.
Step 3: Repeat this procedure until all values for y are found using the algorithm yi ⫽ (ci ⫺ 兺i⫺1
k⫽1 likyk)/lii.
Having determined Y, the solution for the matrix system LTX ⫽ Y is computed in a manner similar to that above. However, this time the solution starts
at the lower right corner and proceeds up the matrix LT. Called a backward
substitution, this is done with the following steps:
Step 1: Compute xn as xn ⫽ yn /lnn.
Step 2: Solve for xn⫺1 as xn⫺1 ⫽ (yn⫺1 ⫺ lmnxn)/l(n⫺1)(n⫺1).
Step 3: Repeat this procedure until all unknowns are computed using the
algorithm xk ⫽ (yk ⫺ 兺nj⫽k⫹1 lkjyj)/lkk.
In this process, once the original values in the constant matrix are accessed
and changed, they are not needed again and the original C matrix can be
overwritten with Y and X matrices so that the entire process requires no
additional storage. In fact, this method of solution also requires fewer oper-
512
COMPUTER OPTIMIZATION
ations than solving X ⫽ (ATWA)⫺1ATWL directly. Table 26.5 lists the computer
codes for these algorithms.
26.6 USING THE CHOLESKY FACTOR TO FIND THE INVERSE OF
THE NORMAL MATRIX
If necessary, the original normal matrix inverse can be found with the Cholesky factor. To derive this matrix, the inverse of the Cholesky factor is computed. The normal matrix inverse is the product of this inverse times its
TABLE 26.5 Computer Algorithms for Forward and Backward Substitutions
BASIC Language:
Rem Forward Substitution
For i ⫽ 1 to Unknown
ix ⫽ VI(i)
C(i) ⫽ C(i) / N(ix⫹i)
For j ⫽ i⫹1 to Unknown
C( j) ⫽ C( j) ⫺ N(ix⫹j) * C(i)
Next j
Next i
Rem Backward Substitution
For i ⫽ Unknown to 1 Step ⫺1
ix ⫽ VI(i)
For j ⫽ Unknown To i⫹1 Step ⫺1
C(i) ⫽ N(ix⫹j) * C( j)
Next j
C(i) ⫽ C(i) / N(ix⫹i)
Next i
FORTRAN Language:
C Forward Substitution
Do 100 i ⫽ 1,Unknown
ix ⫽ VI(i)
C(i) ⫽ C(i) / N(ix⫹i)
Do 100 j ⫽ i⫹1, Unknown
C( j) ⫽ C( j) ⫺ N(ix⫹j) *
C(i)
100 Continue
C Language:
/ / Forward Substitution
for (i⫽1; i⬍⫽unknown; i⫹⫹){
ix ⫽ VI[i];
c[i] ⫽ c[i] / n[ix⫹i];
for ( j⫽i⫹1; j⬍⫽unknown; j⫹⫹){
c[j] ⫽ c[j] ⫺ n[ix⫹j]*c[i];
} / / for j
} / / for i
/ / backward substitution
for (i⫽unknown; i⬎⫽1; i⫺⫺){
ix ⫽ VI[i];
for ( j⫽unknown; j⬎⫽i⫹1; j⫺⫺){
c[i] ⫽ c[i] ⫺ n[ix⫹j] * c[j];
} / / for j
c[i] ⫽ c[i] / n[ix⫹i];
} / / for i
Pascal Language:
{Forward Substitution}
For i :⫽ 1 to Unknown Do Begin
ix :⫽ VI[i];
C[i] :⫽ C[i] / N[ix⫹i];
For j :⫽ i⫹1 to Unknown Do
C[j] :⫽ C[j] ⫺ N[ix⫹j] * C[i];
End; {for i}
C Backward Substitution
Do 110 i ⫽ Unknown, 1,⫺1
ix ⫽ VI(i)
Do 120 j ⫽ Unknown, i⫹1, ⫺1
120 C(i) ⫽ C(i) ⫺ N(ix⫹j) * C( j)
C(i) ⫽ C(i) / N(ix⫹i)
110 Continue
{Backward Substitution}
For i :⫽ Unknown DownTo 1 do Begin
ix :⫽ VI[i];
For j :⫽ Unknown DownTo i⫹1 Do
C[i] :⫽ C[i] ⫺ N[ix⫹j] * C[j];
C[i] :⫽ C[i] / N[ix⫹i];
End; {for i}
26.7
SPARENESS AND OPTIMIZATION OF THE NORMAL MATRIX
513
transpose. The Cholesky factor inverse is determined using the algorithm in
Table 26.6.
Code for this algorithm is shown in Table 26.7.
26.7
SPARENESS AND OPTIMIZATION OF THE NORMAL MATRIX
In the least squares adjustment of most surveying and photogrammetry problems, it is known that certain locations in the normal matrix will contain zeros.
The network shown in Figure 26.2 can be used to demonstrate this fact. In
that figure, assume that distances and angles were observed for every line and
arc, respectively. The following observations are made about the network’s
connectivity. Station 1 is connected to stations 2 and 8 by distance observations and is also connected to stations 2, 3, and 8 by angles. Notice that its
connection to station 3 is due to angles turned at both 2 and 8. Since these
angles directly connect stations 1 and 2, it follows that if the coordinates of
3 change, so will the coordinates of 1, and thus with respect to station 1, the
normal matrix can be expected to have nonzero elements corresponding to
stations 1, 2, 3, and 8. Conversely, the positions corresponding to stations 4,
5, 6, and 7 will have zeros.
Using this analysis with station 3, because it is connected to stations 1, 2,
4, 7, and 8 by angles, zero elements can be expected in the normal matrix
corresponding to stations 5 and 6. Similar analyses can be made for each
station. The resulting symbolic normal matrix representation is shown in Figure 26.3.
A process known as reordering the unknowns can minimize both storage
and computational time. Examine the matrix shown in Figure 26.4, which
results from placing the unknowns of Figure 26.3 in the order 6, 5, 4, 7, 3,
1, 2, and 8. This new matrix has its nonzero elements immediately adjacent
to the diagonal elements, and thus the known zero elements are grouped
together and appear in the leftmost columns of their respective matrix rows.
By modifying the mapping table, storing the known zero elements of the
original matrix in Figure 26.3 can be avoided, as is shown in the column
matrix of Figure 26.4.
This storage scheme requires a mapping table that provides the storage
location for the first nonzero element of each row. Since the diagonal elements
are always nonzero, they can be stored in a separate column matrix without
TABLE 26.6 Pseudocode for an Algorithm Computing the Inverse of a
Cholesky Decomposed Matrix
for i going from the number of unknowns down to 1
for k going from number of unknowns down to i ⫹ 1
for j going from i⫹1 to k, sum the product N(i, j) * N(j, k)
N(i, k) ⫽ ⫺S / N(i, i)
N(i, i) ⫽ 1 / N(i, i)
514
COMPUTER OPTIMIZATION
TABLE 26.7 Computer Algorithms to Find the Inverse of a Cholesky Factored
Matrix
BASIC Language:
{Inverse}
For i ⫽ Unknown To 1 Step ⫺1
ix ⫽ VI(i): ixi ⫽ ix ⫹ i
For k ⫽ Unknown To i⫹1 Step ⫺1
S ⫽ 0!
For j ⫽ i⫹1 To k
S ⫽ S ⫹ N(ix⫹j) * N(VI( j)⫹k)
Next j
N(ix⫹k) ⫽ ⫺S / N(ixi)
Next k
N(ixi) ⫽ 1.0 / N(ixi)
Next i
Fortran Language:
C Inverse
Do 10 i ⫽ Unknown, 1,⫺1
ix ⫽ VI(i)
ixi ⫽ ix ⫹ i
Do 11 k ⫽ Unknown, i⫹1,⫺1
S ⫽ 0.0
Do 12 j ⫽ i⫹1,k
12
S ⫽ S ⫹ N(ix⫹j) * N(VI( j)⫹k)
N(ix⫹k) ⫽ ⫺S / N(ixi)
11 Continue
N(ixi) ⫽ 1.0 / N(ixi)
10 Continue
{Inverse * Transpose of Inverse}
For j ⫽ 1 to Unknown
ixj ⫽ VI( j)
For k ⫽ j to Unknown
S ⫽ 0!
For i ⫽ k to Unknown
S ⫽ S ⫹ N(VI(k)⫹i) *
N(ixj⫹i)
Next i
N(ixj⫹k) ⫽ S
Next k
Next j
C Inverse * Transpose of Inverse}
Do 20 j ⫽ 1, Unknown
ixj ⫽ VI( j)
Do 21 k ⫽ j, Unknown
S ⫽ 0.0
Do 22 i ⫽ k,Unknown
22
S ⫽ S ⫹ N(VI(k)⫹i) * N(ixj⫹i)
N(ixj⫹k) ⫽ S
21 Continue
20 Continue
C Language:
{Inverse}
for (i⫽unknown; i⬎⫽1; i⫺⫺) {
ix ⫽ VI[i]; ixi ⫽ ix ⫹ i;
for(k⫽unknown; k⬎⫽i⫹1; k⫺⫺) {
s ⫽ 0.0;
for ( j⫽i⫹1; j⬍⫽k; j⫹⫹)
s ⫽ s ⫹ n[ix⫹j] * n[VI[j]⫹k];
n[ix⫹k] ⫽ ⫺s / n[ixi];
} / / for k
n[ixi] ⫽ 1.0 / n[ixi];
} / / for i
Pascal Language:
{Inverse}
For i :⫽ Unknown DownTo 1 do Begin
ix :⫽ VI[i]; ixi :⫽ ix ⫹ i;
For k :⫽ Unknown DownTo i⫹1 Do Begin
S :⫽ 0.0;
For j :⫽ i⫹1 To k Do
S :⫽ S ⫹ N[ix⫹j] * N[VI[j]⫹k];
N[ix⫹k] :⫽ ⫺S / N[ixi];
End; {For k}
N[ixi] :⫽ 1.0 / N[ixi];
End; {For i}
{inverse * transpose of inverse}
for ( j⫽1; j⬍⫽unknown; j⫹⫹) {
ixj ⫽ VI[j];
for (k⫽j; k⬍⫽unknown; k⫹⫹) {
s ⫽ 0.0;
For (i⫽k; i⬍⫽unknown; i⫹⫹)
s ⫽ s ⫹ n[VI[k]⫹i] * n[ixj⫹i];
n[ixj⫹k] ⫽ s
} / / for k
} / / for j
{Inverse * Transpose of Inverse}
For j :⫽ 1 to Unknown Do Begin
ixj :⫽ VI[j];
For k :⫽ j to Unknown Do Begin
S :⫽ 0.0;
For i :⫽ k to Unknown Do
S :⫽ S ⫹ N[VI[k]⫹i] * N[ixj⫹i];
N[ixj⫹k] :⫽ S
End; {For k}
End; {For j}
26.7
SPARENESS AND OPTIMIZATION OF THE NORMAL MATRIX
515
Figure 26.2 Series of connected traverses.
loss of efficiency. Ignoring the fact that each station has two unknowns, the
mapping of the first element for each row is 1, 1, 2, 4, 7, 9, 10, 12, and 17.
This mapping scheme enables finding the starting position for each row and
allows for the determination of the off-diagonal length of the row. That is,
row 1 starts at position 1 of the column matrix, but has a length of zero (1
⫺ 1), indicating that there are no off-diagonal elements in this row. Row 2
starts at position 1 of the column matrix and has a length of (2 ⫺ 1) ⫽ 1.
Row 5 starts at 7 and has a length of (9 ⫺ 7) ⫽ 2. By optimizing for matrix
sparseness and symmetry, only 24 (16 ⫹ 8) elements of the original 64element matrix need be stored.
This storage savings, when exploited, can also result in a savings of computational time. To understand this, first examine how the Cholesky factorization procedure processes the normal matrix. For a lower triangular matrix
factorization process, Figure 26.5 shows the manner in which elements are
accessed. Notice that when a particular column is modified, the elements to
the left of this column are used. Also notice that no rows above the corresponding diagonal element are used.
To see how the reorganized sparse matrix can be exploited in the factorization process, the processing steps must be understood. In Figure 26.4 the
Figure 26.3 Normal matrix.
516
COMPUTER OPTIMIZATION
Figure 26.4 Reordered matrix.
first two elements of row 5 are known zeros and each summation loop in the
factorization process should start with the third column. If the zeros in the
off-diagonals are rearranged such that they occur in the leftmost columns of
the rows, it is possible to avoid that portion of the operations for the rows.
Notice how the rows near the lower portion of the matrix are optimized to
minimize the computational effort. A savings of eight multiplications is obtained for computations of column 5. In a large system, the cost savings of
this optimization technique can be enormous. These techniques are discussed
completely by George and Lui (1981). A comparison of the operations performed when computing column 5 of Figure 26.4 in a Cholesky decomposition of both a full and an optimized solution routine are shown in Table
26.8. Note that computations for rows 6 and 7 do not exist since these rows
had known zeros in the columns before column 5 after the reordering process.
If only the more time-consuming multiplication and division operations are
counted, the optimized solution requires five operations compared to 19 in
the nonoptimized (full) solution.
Several methods have been developed to optimize the ordering of the unknowns. Two of the better known reordering schemes are the reverse Cuthill–
Mckee and the banker’s algorithm. Both of these algorithms reorder stations
based on their connectivity. The example of Figure 26.2 will be used to
not accessed
computed and accessed
currently accessed
yet to be accessed
Figure 26.5 Computation of a Cholesky factor.
26.7
SPARENESS AND OPTIMIZATION OF THE NORMAL MATRIX
517
TABLE 26.8 Comparison of the Number of Operations in Computing One
Column
Full Operations
Optimized Operations
1. s ⫽ n512 ⫹ n522 ⫹ 䡠 䡠 䡠 ⫹ n542
2. n55 ⫽ square root of (n55 ⫺ s)
3. j ⫽ 6
(a) s ⫽ n61*n51 ⫹ n62*n52 ⫹ 䡠 䡠 䡠 ⫹
n64*n54
(b) n65 ⫽ (n65 ⫺ s) / n55
j⫽7
(a) s ⫽ n71*n51 ⫹ n72*n52 ⫹ 䡠 䡠 䡠 ⫹
n74*n54
(b) n75 ⫽ (n75 ⫺ s) / n55
j⫽8
(a) s ⫽ n81*n51 ⫹ n82*n52 ⫹ 䡠 䡠 䡠 ⫹
n84*n54
(b) n85 ⫽ (n85 ⫺ s) / n55
s ⫽ n532 ⫹ n542
n55 ⫽ square root of (n55 ⫺ s)
j⫽6
no operations necessary
j⫽7
no operations necessary
j⫽8
s ⫽ n83*n53 ⫹ n84*n54
demonstrate the banker’s algorithm. In this example, a connectivity matrix is
first developed. This is simply a list of stations that are connected to a given
station either by an angle or by distance observations. For instance, station 1
is connected to stations 2 and 8 by distances, and additionally to station 3 by
angles. The complete connectivity matrix of Figure 26.2 is given in the first
two columns of Table 26.9. In addition to the connectivity, the algorithms
also need to know the number of connected points for each station. This is
known as the station’s degree.
To start the reordering, select the station with the lowest degree. In this
example that could be station 1, 2, 5, or 6. For demonstration purposes, begin
with station 6. This station is now removed from each station’s connectivity
matrix, and thus stations 4, 5, and 7 have their degrees reduced by one. This
is shown in column 6. Now select the next station with the smallest degree.
TABLE 26.9 Connectivity Matrix
Station
Connectivity
Degree
6
5
4
7
3
1
2
1
2
3
4
5
6
7
8
2,3,8
1,3,8
1,2,4,7,8
3,5,6,7,8
4,6,7
4,5,7
3,4,5,6,8
1,2,3,4,7
3
3
5
5
3
3
5
5
3
3
5
4
2
—
4
5
3
3
5
3
—
—
3
5
3
3
4
—
—
—
2
4
3
3
3
—
—
—
—
3
2
2
—
—
—
—
—
2
—
1
—
—
—
—
—
1
—
—
—
—
—
—
—
0
518
COMPUTER OPTIMIZATION
Since station 5 had its degree reduced to 2 when station 6 was selected, it
now has the smallest degree. This process is continued; that is, the next station
with the smallest degree is selected and is removed from the connectivity of
the remaining stations. If two or more stations have the smallest degree, the
most recently changed station with the smallest degree is chosen. In this
example, this happened after choosing station 5. Station 4 was selected over
stations 1 and 2 simply because it just had its degree reduced. Although
station 7 qualified equally to station 4, station 4 was higher on the list and
was selected for that reason. This happened again after selecting station 7.
Station 3 was the next choice since it was the first station reduced to a degree
of 3. Note that the final selection of station 8 was omitted from the list.
A similar optimized reordering can be found by starting with station 1, 2,
or 5. This is left as an exercise. For a computer algorithm of this reordering,
the reader should refer to the NOAA Technical Memorandum NGS 4 by
Richard Snay (1976); the computer algorithm for the reverse Cuthill–McKee
is given by George and Lui (1981).
PROBLEMS
26.1
Explain why computer optimization techniques are necessary for doing large least squares adjustments.
26.2
When using double precision, how much computer memory is required in forming the J, W, K, J T, J TW, J TWK, and J TWJ matrices
for a 20-station adjustment having 50 total observations?
26.3
If the normal matrix of Problem 26.2 were stored in lower triangular
form and computed directly from the observations, how much storage
would be required?
26.4
Discuss how computational savings are created when solving the normal equations by Cholesky factorization and substitution.
26.5
Write a comparison of an operations table similar to that in Table
26.8 for column 3 of the matrix shown in Figure 26.4.
26.6
Repeat Problem 26.5 for column 6.
26.7
Using the station reordering algorithm presented in Section 26.7, develop a connectivity matrix and reorder the stations when starting
with station 1. Draw the normal matrix of the newly reordered station,
which is similar to the sketch shown in Figure 26.4.
26.8
Repeat Problem 26.7, starting the reordering with station 2.
26.9
Repeat Problem 26.7, starting the reordering with station 5.
PROBLEMS
519
Programming Problems
26.10 Develop a least squares program similar to those developed in Chapter
16 that builds the normal equations directly from the observations and
uses a Cholesky factorization procedure to find the solution.
26.11 Develop a Mathcad worksheet that will solve Problem 26.10.
APPENDIX A
INTRODUCTION TO MATRICES
A.1
INTRODUCTION
Matrix algebra provides at least two important advantages: (1) it enables reducing complicated systems of equations to simple expressions that can be
visualized and manipulated more easily, and (2) it provides a systematic,
mathematical method for solving problems that is well adapted to computers.
Problems are frequently encountered in surveying, geodesy, and photogrammetry that require the solution of large systems of equations. This book deals
specifically with the analysis and adjustment of redundant measurements that
must satisfy certain geometric conditions. This frequently results in large
equation systems which, when solved according to the least squares method,
yield most probable estimates for adjusted observations and unknown parameters. As will be demonstrated, matrix methods are particularly well suited
for least squares computations, and in this book they are used for analyzing
and solving these equation systems.
A.2
DEFINITION OF A MATRIX
A matrix is a set of numbers or symbols arranged in a square or rectangular
array of m rows and n columns. The arrangement is such that certain defined
mathematical operations can be performed in a systematic and efficient manner. As an example of a matrix representation, consider the following system
of three linear equations involving three unknowns:
520
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
A.3
SIZE OR DIMENSIONS OF A MATRIX
521
a11x1 ⫹ a12 x2 ⫹ a13 x3 ⫽ c1
a21x1 ⫹ a22 x2 ⫹ a23 x3 ⫽ c2
(A.1)
a31x1 ⫹ a32 x2 ⫹ a33 x3 ⫽ c3
In Equation (A.1), the a’s are coefficients of the unknowns x’s, and the c’s
are the constant terms. The system above can be represented in summation
notation as
冘a x ⫽c
3
1i i
1
i⫽1
冘a x ⫽c
3
2i i
2
i⫽1
冘a x ⫽c
3
3i i
3
i⫽1
It can also be represented in matrix form as
冤
a11 a12
a21 a22
a31 a32
冥冤 冥 冤 冥
a13
a23
a33
x1
c1
x2 ⫽ c2
x3
c3
(A.2)
In turn, Equation (A.2) can be represented in compact matrix notation as
AX ⫽ C
(A.3)
In Equation (A.3), capital letters (A, X, and C) are used to denote a matrix
or an array of numbers or symbols. From simplified equation (A.3), it is
immediately obvious that matrix methods provide a compact shorthand notation convenient for handling a large system of equations.
A.3
SIZE OR DIMENSIONS OF A MATRIX
The size or dimension of a matrix is specified by its number of rows m and
its number of columns n. Thus, matrix 2D3 below is a 2 ⫻ 3 matrix; that is,
there are two rows and three columns (i.e., m ⫽ 2, n ⫽ 3). Notice that the
subscript indicates the number of rows and the superscript indicates the number of columns in the notation 2D3.
522
INTRODUCTION TO MATRICES
2
D3 ⫽
冋
d11 d12
d21 d22
册 冋
d13
3
⫽
d23
5
2
4
册
6
1
Also, the matrix
冤
冥 冤 冥
e11 e12
7 1
2
E
⫽
e
e
⫽
4
3
3
21
22
e31 e32
2 8
is a 3 ⫻ 2 matrix.
Note that the position of an element in a matrix is defined by a double
subscript and that a lowercase letter is used to designate any particular element within a matrix. Thus, d23 ⫽ 1 is in row 2 and column 3 of the D matrix
above. In general, the subscript ij indicates an element’s position in a matrix,
where i represents the row and j the column.
A.4
TYPES OF MATRICES
Several different types of matrices exist as described below. Various symbols
can be used to designate them, as illustrated.
1. Column matrix. The number of rows can be any positive integer, but
the number of columns is 1.
冤 冥
3
A ⫽ ⫺2
5
2. Row matrix. The number of columns can be any positive integer, but
the number of rows is 1.
A ⫽ [6 ⫺4
2]
3. Rectangular matrix. The number of rows and columns are m and n,
respectively, where m and n are any positive integers.
A⫽
冋
册
3 2 6
5 4 1
4. Square matrix. The number of rows equals the number of columns.
冤
冥
4
2 ⫺5
A ⫽ ⫺7
3
4
6 ⫺1
9
A.5
MATRIX EQUALITY
523
A square matrix, for which the determinant is zero, is termed singular.
If the determinant is nonzero, it is termed nonsingular. (The determinant
of a matrix is discussed in Section B.3.)
5. Symmetric matrix. The matrix is mirrored about the main diagonal going
from top left to bottom right (i.e., element aij ⫽ element aji). A symmetric matrix is always a square matrix.
冤
冥
2 ⫺4 6
A ⫽ ⫺4
7 3
6
3 5
6. Diagonal matrix. Only the elements on the main diagonal are not zero.
The diagonal matrix is always a square matrix.
冤
7
0 0
A ⫽ 0 ⫺3 0
0
0 6
冥
7. Unit matrix. This is a diagonal matrix with 1’s along the main diagonal.
It is also called an identity matrix and is usually identified by the symbol I.
冤
冥
1 0 0
I⫽ 0 1 0
0 0 1
8. Transpose of a matrix. This is obtained by interchanging rows and columns (i.e., element aTij ⫽ element aji). Thus, the dimensions of AT are
the reverse of the dimensions of A. If
A⫽
冋
册
2 4 7
5 3 1
the transpose of A, denoted AT, is
冤 冥
2
AT ⫽ 4
7
A.5
5
3
1
MATRIX EQUALITY
Two matrices are said to be equal only when they are equal element by
element. Thus, the two matrices must be the same size or have the same
dimensions.
524
INTRODUCTION TO MATRICES
A⫽
A.6
冋
册
冋
册
1 7 6
1 7 6
⫽B⫽
4 3 2
4 3 2
ADDITION OR SUBTRACTION OF MATRICES
Matrices can be added or subtracted, but to do so, they must have the same
dimensions. If two matrices have equal dimensions, they are said to be conformable for addition or subtraction. In adding or subtracting matrices, elements from each unique row/column position of the two matrices are added
or subtracted systematically and the sum or difference is placed in the same
unique row/column location of the resulting matrix. The following example
illustrates this procedure.
2
A3 ⫹ 2B3 ⫽
冋
7
2
3
⫺1
⫺5
6
册 冋
⫹
1
5
⫺4
⫺2
册 冋
册
6
8
8 5
⫽
⫽ 2C 3
3
⫺2 ⫺7 9
Assuming that two matrices are conformable for addition or subtraction,
the following are true:
(a) A ⫹ B ⫽ B ⫹ A
(b) A ⫹ (B ⫹ C) ⫽ (A ⫹ B) ⫹ C
A.7
(commutative law)
(associative law)
SCALAR MULTIPLICATION OF A MATRIX
Matrices can be multiplied by a scalar (i.e., a constant). Let k be any scalar
quantity; then
kA ⫽ Ak
The following are examples.
冤 冥冤 冥 冤 冥
3
2
4⫻
4
5
⫺1
3 ⫺1
12 ⫺4
6
2
6
8 24
⫽
⫻4⫽
7
4
7
16 28
3
5
3
20 12
As illustrated above, each element of the matrix A is multiplied by the scalar
k to obtain the elements of C. Note that 4 ⫻ A ⫽ A ⫹ A ⫹ A ⫹ A ⫽ A ⫻
4.
A.8
A.8
MATRIX MULTIPLICATION
525
MATRIX MULTIPLICATION
If matrix A is to be postmultiplied by matrix B (i.e., the product of A ⫻ B
determined), the number of columns in matrix A must equal the number of
rows in matrix B. This is a basic requirement for matrix multiplication. When
this condition is satisfied, A and B are said to be conformable for multiplication. The product C will have the same number of rows as A and the same
number of columns as B. Thus, the following multiplications are possible:
4
A2 ⫻ 2B3 ⫽ 4C3
1
A3 ⫻ 3B1 ⫽ 1C1
3
A3 ⫻ 3B1 ⫽ 3C1
These multiplications are not possible:
2
B3 ⫻ 4A2
6
A2 ⫻ 6B3
To demonstrate the process of matrix multiplication, consider the following
example:
2
A3 ⫻ 3B2 ⫽
冋
a11
a21
a12
a22
册冤
a13
a12
冥 冋
b11 b12
c
b21 b22 ⫽ 11
c21
b31 b32
册
c12
⫽ 2C2
c22
The elements cij of matrix C are the total sums obtained by multiplying
each element in row i of matrix A successively by the elements in column j
of matrix B and then summing these products. Thus, for the example above,
冘a b
3
c11 ⫽ a11 b11 ⫹ a12 b21 ⫹ a13 b31 ⫽
1i i1
i⫽1
冘a b
3
c12 ⫽ a11 b12 ⫹ a12 b22 ⫹ a13 b32 ⫽
1i i2
i⫽1
冘a b
3
c21 ⫽ a21 b11 ⫹ a22 b21 ⫹ a23 b31 ⫽
2i i1
i⫽1
冘a b
3
c22 ⫽ a21 b12 ⫹ a22 b22 ⫹ a23 b32 ⫽
2i i2
i⫽1
The process above is seen more easily with a numerical example.
526
INTRODUCTION TO MATRICES
冋
册冤 冥 冋
1 2 3
4 2 7
册 冋
4 8
c
c
31 21
6 2 ⫽ 11 12 ⫽
c21 c22
63 57
5 3
册
c11 ⫽ 1 ⫻ 4 ⫹ 2 ⫻ 6 ⫹ 3 ⫻ 5 ⫽ 31
c12 ⫽ 1 ⫻ 8 ⫹ 2 ⫻ 2 ⫹ 3 ⫻ 3 ⫽ 21
c21 ⫽ 4 ⫻ 4 ⫹ 2 ⫻ 6 ⫹ 7 ⫻ 5 ⫽ 63
c22 ⫽ 4 ⫻ 8 ⫹ 2 ⫻ 2 ⫹ 7 ⫻ 3 ⫽ 57
The student should now verify the matrix representation of Equations (A.1)
and (A.2). Notice that the product of a unit matrix, I, and a conformable
matrix, A (one with the same number of rows as I), equals the original matrix
A. Thus,
2
I 2 ⫻ 2 A2 ⫽
冋 册冋 册 冋 册
1 0
0 1
5 6
5 6
⫽
⫽ 2 A2
7 8
7 8
Assuming that matrices A, B, and C are conformable for multiplication and
in the order indicated, the following are true:
(c)
(d)
(e)
(f)
A(B⫹C) ⫽ AB ⫹ AC
(A⫹B)C ⫽ AC ⫹ BC
A(BC) ⫽ (AB)C
(AB)T ⫽ BTAT
(first distributive law)
(second distributive law)
(associative law)
The following cautions are also stated:
(g) AB is not generally equal to BA, and BA may not even be conformable.
(h) If AB ⫽ 0, neither A nor B necessarily ⫽ 0.
(i) If AB ⫽ AC, B does not necessarily ⫽ C.
Let A ⫽
Example of (c):
冋 册
2 1
1 2
B⫽
冋
册
⫺1
⫺1
2
2
C⫽
冋 册
1 1
1 1
A.8
A(B ⫹ C) ⫽
⫽
527
MATRIX MULTIPLICATION
冋 册 冋 册
冋 册 冋 册 冋 册
2 1
3 3
⫻
⫽ AB ⫹ AC
1 2
6 6
0
3
0
3
⫹
3
3
3
3
⫽
3
6
3
6
冋 册冋 册 冋 册
冋 册 冋 册 冋 册
Example of (d)1:
1 0
3 4
(A ⫹ B)C ⫽
3
3
⫽
1 1
1 1
⫽
⫽ AC ⫹ BC
1 1
7 7
3
⫺2 ⫺2
1 1
⫹
⫽
3
4
4
7 7
Example of (e):
A(BC) ⫽
⫽
冋 册冋 册 冋 册
冋 册冋 册 冋 册
2 1
1 2
⫺2
⫺2
4
4
0 0
3 3
1 1
0 0
⫽
1 1
6 6
⫽
0 0
⫽ (AB)C
6 6
Example of (f):
冋
册
2 6 4
1 2 7
Let A ⫽
冤 冥
3 2
and B ⫽ 9 0
1 3
冋
册
册冤 冥 冋
册
then AB ⫽
64 16
28 23
so
(AB)T ⫽
冋
64 28
16 23
册
and BT AT ⫽
冋
2 1
64 28
6 2 ⫽
16 23
4 7
3 9 1
2 0 3
Example of (g):
AB ⫽
冋 册 冋
册
0 0
⫺3 ⫺3
⫽
⫽ BA
3 3
6
6
Example of (h):
Let A ⫽
冋 册
2 2
1 1
and
B⫽
冋
⫺1
⫺2
1
2
册
Then AB ⫽ 0, but neither A nor B equal 0.
1
The multiplication symbol, ⫻, is generally not written in matrix equations. This convention of
not using ⫻ is followed in this book.
528
INTRODUCTION TO MATRICES
Example of (i):
Let A ⫽
冋 册
2
1
冋 册
2
1
, B⫽
1
1
2
,
3
and C ⫽
冋
册
4
10
⫺2
⫺5
Then AB ⫽ AC, but B ⫽ C, where
AB ⫽
A.9
冋 册
4
2
10
⫽ AC
5
COMPUTER ALGORITHMS FOR MATRIX OPERATIONS
It should be apparent that addition, subtraction, and multiplication of large
matrices involves many arithmetic operations. These are very tedious when
done by hand but can be done quickly by a computer. In this section, general
mathematical expressions are developed for performing these operations using
a computer. These general mathematical expressions, when programmed for
computer solution, are called algorithms.
A.9.1
Addition or Subtraction of Two Matrices
Note that the two matrices A and B in Figure A.1 are conformable for addition. Find the sum of the two matrices and place the results in C.
Step 1: Add the first element (a11) of the A matrix to the first element (b11)
of B, placing the result in the first element (c11) of C. Repeat this process
for each of the successive columns along the first row of A and B, or from
j ⫽ 1 to j ⫽ n.
Step 2: Iterate step 1 for each successive row of the matrices, or for rows
increasing from i ⫽ 1 to i ⫽ m. Table A.1 shows this entire operation of
adding two matrices in the four computer languages BASIC, C, FORTRAN, and Pascal.
Figure A.1 Addition of matrices.
A.9
COMPUTER ALGORITHMS FOR MATRIX OPERATIONS
529
TABLE A.1 Addition Algorithm in BASIC, C, FORTRAN, and Pascal
BASIC Language:
1000 For i ⫽ 1 to M
1010 For j ⫽ 1 to N
1020
C(i,j) ⫽ A(i,j) ⫹ B(i,j)
1030
Next j
1040 Next i
FORTRAN Language:
Do 100 I ⫽ 1,M
Do 100 J ⫽ 1,N
C(i,j) ⫽ A(i,j) ⫹ B(i,j)
100 Continue
C Language:
for (i⫽0; i⬍m; i⫹⫹)
for (j⫽0; j⬍n; j⫹⫹)
C[i][j] ⫽ A[i][j] ⫹ B[i][j];
Pascal Language:
For i :⫽ 1 to M do
For j :⫽ 1 to N do
C[i,j] :⫽ A[i,j] ⫹ B[i,j];
A.9.2
Matrix Multiplication
Consider the two matrices A and B in Figure A.2, which are conformable for
multiplication. Find the product AB and place the results in C.
Step 1: Sum the products (A row i ⫽ 1) ⫻ (B column k ⫽ 1), with the
elements of A and B being increased successively from j ⫽ 1 to j ⫽ p.
Place the result in cik, or c11, for this first step. Mathematically, this step
is represented as
冘ab
p
cik ⫽
ij jk
with i ⫽ 1 and k ⫽ 1
j⫽1
Step 2: Increase k successively by 1, repeat step 1, and place the results in
cik. Continue increasing k and repeating step 1 until k ⫽ p.
Step 3: Increase i from 1 to 2, and repeat steps 1 and 2 in entirety. Upon
completion with i ⫽ 2, increment i by 1 and repeat steps 1 and 2 again in
entirety. Continue this process through i ⫽ m. This completes the matrix
multiplication.
Figure A.2 Multiplication of matrices.
530
TABLE A.2 Multiplication Algorithm in BASIC, C, FORTRAN, and Pascal
BASIC Language:
For i ⫽ 1 to M
For k ⫽ 1 to N
C(i,k) ⫽ 0.0
For j ⫽ 1 to P
C(i,k) ⫽ C(i,k) ⫹ A(i,j)*B(j,k)
Next j : Next k: Next I
FORTRAN Language:
Do 100 i ⫽ 1,M
Do 100 k ⫽ 1,N
C(i,k) ⫽ 0.0
Do 100 j ⫽ 1,P
C(i,k)⫽C(i,k)⫹A(i,j)*B(j,k)
100 Continue
C Language:
for (i⫽0; i⬍m; i⫹⫹)
for (k⫽0; k⬍n; k⫹⫹) {
C[i][k] ⫽ 0;
for (j⫽0; j⬍p; j⫹⫹)
C[i][k] ⫽ C[i][k] ⫹ A[i][j]*B[j][k];
} / / for k
Pascal Language:
For i :⫽ 1 to M do
For k :⫽ 1 to N do Begin
C[i,k] :⫽ 0.0;
For j :⫽ 1 to P do
C[i,k] :⫽ C[i,k] ⫹ A[i,j]*B[j,k]
End; {for k}
PROBLEMS
531
This operation is shown Table A.2 in the four languages BASIC, C, FORTRAN, and Pascal. As a final note on multiplication, it is essential that the
order in which the matrices are multiplied be specified.
A.10
USE OF THE MATRIX SOFTWARE
A software package called MATRIX is included on the CD provided with the
book. It includes all the matrix operations that will be necessary to study the
subject of adjustment computations and solve the after-chapter homework
problems herein that require matrix manipulation. Instructions for use of the
software are given in Appendix F.
PROBLEMS
A.1
Suppose that the following system of linear equations are to be represented in matrix form as AX ⫽ B. Formulate the three matrices.
2x1 ⫹ x2 ⫹ 5x3 ⫹ x4 ⫽
5
x1 ⫹ x2 ⫺ 3x3 ⫹ 4x4 ⫽ ⫺1
3x1 ⫹ 6x2 ⫺ 2x3 ⫹ x4 ⫽
8
2x1 ⫹ 2x2 ⫹ 2x3 ⫺ 3x4 ⫽
2
A.2
For Problem A.1, find the product of ATA.
A.3
Do the following matrix operations.
(a)
(b)
冤
冤
冤
1
2
4
0
2 ⫺5
1
4
2
冥 冤
冥 冤
冥
册冤 冥
⫺1
⫹
3
1
2
2 ⫺1 0
0
2 1
⫺5
1 2
⫺
3 ⫺4 1
2
1
5 0
3
2 ⫺2 3 ⫺1
1
2 ⫺1 0
(c) 3 4
0
2 1
2 ⫺5
1 2
(d)
冋
1 2 1
4 0 2
3 ⫺4
1
5
⫺2
2
⫺4
1
2
5 0 ⫺3
⫺2 3 ⫺1
冥
冥
0
2 1
1 2
532
INTRODUCTION TO MATRICES
A.4
Solve Problem A.2 using the MATRIX software.
A.5
Solve Problem A.3 using the MATRIX software.
A.6
Let
A⫽
冋 册
2 0
3 1
冋 册
4 ⫺1
0
2
B⫽
C⫽
冋 册
1 0
0 1
D⫽
冋 册
0 0
0 0
(a) Find AB.
(b) Find B2.
(c) Find CB.
(d) Find BA.
(e) Find B3.
(f) Find DB.
(g) Find C3.
A.7
Do Problem A.6 using the MATRIX software.
A.8
Multiply the following, if possible. If not possible, give reasons why
the multiplication cannot be done.
(a)
冋 册
2 1
[1
4 0
(b) [3
(c)
(d)
A.9
1
2]
冤冥
4
3] 0
9
冥冤
冤
2 3
4 4
1 0 ⫺1 6
0 1
2 9
冋册
0
2
3
1
1
0
0 ⫺1
冥
2
[3 ⫺1]
0
Expand the following summations into their equivalent algebraic expressions.
冘k
(b) 冘 (k ⫺ 2)
(c) 冘 a
(d) 冘 a a
5
(a)
k⫽1
7
k⫽3
4
k
k⫽1
3
2k k3
k⫽1
PROBLEMS
533
冘 冘 (a ⫹ b )
(f) 冘 冘 冘 a b
3
4
(e)
ij
ij
i⫽1 j⫽1
2
2
3
ik kj
i⫽1 j⫽1 k⫽1
A.10
Write the algebraic expressions for the system of equations represented by the following matrix notation.
冤 冥冋 册 冤 冥
2 0
1 3
4 2
A.11
2
x1
⫽ 1
x2
3
Show that, in general, (AB)T ⫽ BTAT.
Programming Problems
A.12
Write a program that will read and write the elements of a matrix,
compute the transpose of the matrix, and write the solution. (Hint:
Place the reading, writing, and transposition codes in separate
subroutines/procedures/functions to be called from the main program.)
A.13
Select one of the matrix addition codes from Table A.1 and enter it
and any necessary supporting code to solve Problem A.3(a). (Hint:
Place the code in Table A.1 in a separate subroutine/procedure/function to be called from the main program.)
A.14
Select one of the matrix multiplication programs from Table A.2, and
enter this code and any necessary supporting code to solve Problem
A.3(d). (Hint: Place the code in Table A.2 in a separate subroutine/
procedure/function to be called from the main program.)
A.15
Write a program that will read and write a file of matrices, and do the
matrix operations of transposition, addition, subtraction, and multiplication. Demonstrate the program by doing Problems A.2 and A.3 (a),
(b), and (d). (Hint: Place the reading, writing, transposition, addition, subtraction, and multiplication codes in separate subroutines/
procedures/functions to be called from the main program.)
A.16
Write a Mathcad worksheet that solves Problems A.3 and A.6.
APPENDIX B
SOLUTION OF EQUATIONS
BY MATRIX METHODS
B.1
INTRODUCTION
As stated in Appendix A, an advantage offered by matrix algebra is its adaptability to computer use. Using matrix algebra, large systems of simultaneous
linear equations can be programmed for general computer solution using only
a few systematic steps. The simplicities of programming matrix additions and
multiplications, for example, were presented in Section A.9. To solve a system
of equations using matrix methods, it is first necessary to define and compute
the inverse matrix.
B.2
INVERSE MATRIX
If a square matrix is nonsingular (its determinant is not zero), it possesses an
inverse matrix. When a system of simultaneous linear equations consisting of
n equations and involving n unknowns is expressed as AX ⫽ B, the coefficient
matrix (A) is a square matrix of dimensions n ⫻ n. Consider the system of
linear equations
AX ⫽ B
(B.1)
The inverse of matrix A, symbolized as A⫺1, is defined as
A⫺1A ⫽ I
(B.2)
where I is the identity matrix. Premultiplying both sides of matrix Equation
(B.1) by A⫺1 gives
534
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
B.3
INVERSE OF A 2 ⫻ 2 MATRIX
535
A⫺1AX ⫽ A⫺1B
Reducing yields
IX ⫽ A⫺1B
X ⫽ A⫺1B
(B.3)
Thus, the inverse is used to find the matrix of unknowns, X. The following
points should be emphasized regarding matrix inversions:
1. Square matrices have inverses, with the exception noted below.
2. When the determinant of a matrix is zero, the matrix is said to be
singular and its inverse cannot be found.
3. The inversion of even a small matrix is a tedious and time-consuming
operation when done by hand. However, when done by a computer, the
inverse can be found quickly and easily.
B.3
INVERSE OF A 2 ⫻ 2 MATRIX
Several general methods are available for finding a matrix inverse. Two are
considered herein. Before proceeding with general cases, however, consider
the specific case of finding the inverse for a 2 ⫻ 2 matrix using simple
elementary matrix operations. Let any 2 ⫻ 2 matrix be symbolized as A.
Also, let
冋 册
a b
c d
A⫽
and A⫺1 ⫽
冋 册
w x
y z
By applying Equation (B.2) and recalling the definition of an identity matrix
I as given in Section A.4, it is possible to calculate w, x, y, and z in terms of
a, b, c, and d of A⫺1. Substituting in the appropriate values gives
冋 册冋 册 冋 册
a b
c d
w x
1 0
⫽
y z
0 1
By matrix multiplication
aw ⫹ by ⫽ 1
(a)
ax ⫹ bz ⫽ 0
(b)
cw ⫹ dy ⫽ 0
(c)
cx ⫹ dz ⫽ 1
(d)
The determinant of A is symbolized as 㛳 A㛳 and is equal to ad ⫺ bc.
536
SOLUTION OF EQUATIONS BY MATRIX METHODS
冦
冦
冦
冦
冧
冧
冧
冧
1 ⫺ aw
cw
; from (b) y ⫽ ⫺
b
d
1 ⫺ aw
cw
d
d
then
⫽⫺
; reducing: w ⫽
⫽
b
d
da ⫺ bc 㛳 A㛳
From (a) y ⫽
ax
1 ⫺ cx
; from (d) z ⫽
b
d
ax 1 ⫺ cx
b
⫺b
then ⫺
⫽
; reducing: x ⫽
⫽
b
d
⫺da ⫹ bc
㛳 A㛳
From (b) z ⫽ ⫺
1 ⫺ by
dy
; from (c) w ⫽ ⫺
a
c
1 ⫺ by
dy
c
⫺c
then ⫺
⫽ ⫺ ; reducing: y ⫽
⫽
a
c
⫺ad ⫹ cb
㛳 A㛳
From (a) w ⫽
bz
1 ⫺ dz
; from (d) x ⫽
a
c
bz 1 ⫺ dz
a
a
then ⫺
⫽
; reducing: z ⫽
⫽
a
c
ad ⫺ bc 㛳 A㛳
From (b) x ⫽ ⫺
Thus, for any 2 ⫻ 2 matrix composed of the elements
冋 册
a b
, its inverse is
c d
simply
冤 冥
d
Example B.1 If A ⫽
㛳 A㛳
⫺b
㛳 A㛳
⫺b
㛳 A㛳
a
㛳 A㛳
⫽
1
㛳 A㛳
冋
d
⫺b
⫺b
a
册
冋 册
2
4
3
, find A⫺1.
1
SOLUTION
A⫺1 ⫽ ⫺
冋
1
1
10 ⫺4
册
⫺3
2
A check on the inverse can be obtained by testing it against its definition, or
A⫺1A ⫽ I. Thus,
B.4
⫺
B.4
冋
1
1
10 ⫺4
INVERSES BY ADJOINTS
537
册冋 册 冋 册
⫺3
2
2 3
1 0
⫽
⫽I
4 1
0 1
INVERSES BY ADJOINTS
The inverse of A can be found using the method of adjoints with the following
equation:
A⫺1 ⫽
adjoint of A
adjoint of A
⫽
determinant of A
㛳 A㛳
(B.4)
The adjoint of A is obtained by first replacing each matrix element by its
signed minor or cofactor, and then transposing the resulting matrix. The cofactor of element aij equals (⫺1)i⫹j times the numerical value of the determinant for the remaining elements after row i and column j have been
removed from the matrix. This procedure is illustrated in Figure B.1, where
the cofactor of a12 is
(⫺1)1⫹2 (a21a33 ⫺ a31a23) ⫽ a31a23 ⫺ a21a33
Using this procedure, the inverse of the following A matrix is found:
冤
冥 冤
a11 a12 a13
4 3 2
A ⫽ a21 a22 a23 ⫽ 3 4 1
a31 a32 a33
2 3 4
冥
For this A matrix, the cofactors are calculated as follows
Figure B.1 Cofactor of the a12 element.
538
SOLUTION OF EQUATIONS BY MATRIX METHODS
cofactor of a11 ⫽ (⫺1)2 (4 ⫻ 4 ⫺ 1 ⫻ 3) ⫽ 13
cofactor of a21 ⫽ (⫺1)3 (3 ⫻ 4 ⫺ 2 ⫻ 3) ⫽ ⫺6
cofactor of a31 ⫽ (⫺1)4 (3 ⫻ 1 ⫺ 2 ⫻ 4) ⫽ ⫺5
cofactor of a12 ⫽ (⫺l)3 (3 ⫻ 4 ⫺ 1 ⫻ 2) ⫽ ⫺10
Following the procedure above, the matrix of cofactors is
冤
冥
13 ⫺10
1
matrix of cofactors ⫽ ⫺6
12 ⫺6
⫺5
2
7
Transposing this cofactor matrix produces the following adjoint of A:
冤
13 ⫺6
adjoint of A ⫽ ⫺10 12
1 ⫺6
⫺5
2
7
冥
The determinant of A is the sum of the products of the elements in the
first row of the original matrix times their respective cofactors. Since the
cofactors were obtained in the previous step, this simplifies to
㛳 A㛳 ⫽ 4(13) ⫹ 3(⫺10) ⫹ 2(1) ⫽ 24
The inverse of A is now calculated as
A⫺1 ⫽
冤
冥 冤
冥
13 ⫺6 ⫺5
13/24 ⫺1/4 ⫺5/24
1
⫺10 12
2 ⫽ ⫺5/12
1/2
1/12
24
1 ⫺6
7
1/24 ⫺1/4
7/24
Again, a check on the arithmetical work is obtained by using the definition
of an inverse:
冤
冥冤
13/24 ⫺1/4 ⫺5/24
AA⫺1 ⫽ ⫺5/12
1/2
1/12
1/24 ⫺1/4
7/24
B.5
冥 冤
冥
4 3 2
1 0 0
3 4 1 ⫽ 0 1 0 ⫽I
2 3 4
0 0 1
INVERSES BY ROW TRANSFORMATIONS
1. The multiplication of every element in any row by a nonzero scalar.
2. The addition (or subtraction) of the elements in any row to the elements
of any other row.
3. Combinations of 1 and 2.
B.5
INVERSES BY ROW TRANSFORMATIONS
539
If elementary row transformations are performed successively on A such
that A is transformed into I, and if throughout the procedure the same row
transformations are also done to the same rows of the identity matrix I, the I
matrix will be transformed into A⫺1. This procedure is illustrated using the
same matrix as that used to demonstrate the method of adjoints.
Initially, the original matrix and the identity matrix are listed side by side:
A
I
4 3 2 兩 1 0 0
3 4 1 兩 0 1 0
2 3 4 兩 0 0 1
冤
冥
With the following three row transformations performed on A and I, they are
transformed into matrices A1 and I1, respectively:
1. Multiply row 1 of matrices A and I by 1/a11, or 1/4. Place the results
in row 1 of A1 and I1, respectively. This converts a11 of matrix A1 to 1,
as shown below.
冤
1
3
2
冥
3/4 1/2 兩 1/4 0 0
4
1 兩
0 1 0
3
4 兩
0 0 1
2. Multiply row 1 of matrices A1 and I1 by a21, or 3. Subtract the resulting
row from row 2 of matrices A and I and place the difference in row 2
of A1 and I1, respectively. This converts a21 of A1 to zero.
3. Multiply row 1 of matrices A1 and I1 by a31, or 2. Subtract the resulting
row from row 3 of matrices A and I and place the difference in row 3
of A1 and I1, respectively. This changes a31 of A1 to zero.
After doing these operations, the transformed matrices A1 and I1 are
A1
I1
1 3/4
1/2 兩
1/4 0 0
0 7/4 ⫺1/2 兩 ⫺3/4 1 0
0 3/2
3 兩 ⫺1/2 0 1
冤
冥
Notice that the first column of A1 has been made equivalent to the first
column of a 3 ⫻ 3 identity matrix as a result of these three row transformations. For matrices having more than three rows, this same general procedure would be followed for each row to convert the first element in each
row of A1 to zero, with the exception to first row of A.
Next, the following three elementary row transformations are done on matrices A1 and I1 to transform them into matrices A2 and I2:
540
SOLUTION OF EQUATIONS BY MATRIX METHODS
1. Multiply row 2 of A1 and I1 by 1/a22, or 4/7, and place the results in
row 2 of A2 and I2. This converts a22 to 1, as shown below.
冤
冥
1 3/4
1/2 兩
1/4
0 0
0
1 ⫺2/7 兩 ⫺3/7 4/7 0
0 3/2
3 兩 ⫺1/2
0 1
2. Multiply row 2 of A2
from row 1 of A1 and
respectively.
3. Multiply row 2 of A2
from row 3 of A1 and
respectively.
and I2 by a12, or 3/4. Subtract the resulting row
I1 and place the difference in row 1 of A2 and I2,
and I2 by a32, or 3/2. Subtract the resulting row
I1 and place the difference in row 3 of A2 and I2,
After doing these operations, the transformed matrices A2 and I2 are
A2
I2
1 0
5/7 兩
4/7 ⫺3/7 0
0 1 ⫺2/7 兩 ⫺3/7
4/7 0
0 0 24/7 兩
1/7 ⫺6/7 1
冤
冥
Notice that after this second series of steps is completed, the second column
of A2 conforms to column two of a 3 ⫻ 3 identity matrix. Again, for matrices
having more than three rows, this same general procedure would be followed
for each row, to convert the second element in each row (except the second
row) of A2 to zero.
Finally, the following three row transformations are applied to matrices A2
and I2 to transform them into matrices A3 and I3. These three steps are:
1. Multiply row 3 of A2 and I2 by 1/a33, or 7/24 and place the results in
row 3 of A3 and I3, respectively. This converts a33 to 1, as shown below.
冤
冥
1 0
5/7 兩
4/7 ⫺3 /7
0
0 1 ⫺2/7 兩 ⫺3/7
4/7
0
0 0
1 兩 1/24 ⫺1/4 7/24
2. Multiply row 3 of A2 and I2 by a13, or 5/7. Subtract the results from
row 1 of A2 and I2 and place the difference in row 1 of A3 and I3,
respectively.
3. Multiply row 3 of A2 and I2 by a23, or ⫺2/7. Subtract the results from
row 2 of A2 and I2 and place the difference in row 2 of A3 and I3.
Following these operations, the transformed matrices A3 and I3 are
TABLE B.1 Inverse Algorithm in BASIC, C, FORTRAN, and PASCAL
541
BASIC Language:
REM INVERT A MATRIX
FOR k ⫽ 1 TO n
FOR j ⫽ 1 TO n
IF j⬍⬎k THEN A(k,j) ⫽ A(k,j) / A(k,k)
NEXT j
A(k,k) ⫽ 1 / A(k,k)
FOR i ⫽ 1 TO n
IF i⬍⬎k THEN
FOR j⫽1 TO n
IF j⬍⬎k THEN A(i,j) ⫽ A(i,j) ⫺ A(i,k)*A(k,j)
NEXT j
A(i,k) ⫽ ⫺A(i,k)*A(k,k)
END IF
NEXT i: NEXT k
C Language:
for (k⫽0; k⬍n; k⫹⫹) {
for (j⫽0; j⬍n; j⫹⫹)
if (j!⫽k) A[k][j] ⫽ A[k][j] / A[k][k];
A[k][k] ⫽ 1.0 / A[k][k];
for (i⫽0; i⬍n; i⫹⫹)
if (i!⫽k) {
for (j⫽0; j⬍n; j⫹⫹)
if (j!⫽k) A[i][j] ⫽ A[i][j] ⫺ A[i][k]*A[k][j];
A[i][k] ⫽ ⫺A[i][k]*A[k][k];
} / / if i⬍⬎k
} / / for k
FORTRAN Language:
Do 560 k ⫽ 1,N
Do 520 j ⫽ 1,N
If (j.NE.k) Then
A(k,j) ⫽ A(k,j) / A(k,k)
520 Continue
A(K,K) ⫽ 1.0 / A(K,K)
Do 560 i ⫽ 1,N
If (i.EQ.k) Then GOTO 560
Do 550 j ⫽ 1,N
If (j.NE.k) Then
A(i,j) ⫽ A(i,j) ⫺ A(i,k)*A(k,j)
550 Continue
A(i,k) ⫽ ⫺A(i,k) * A(k,k)
560 Continue
Pascal Language:
For k :⫽ 1 to N do Begin
For j :⫽ 1 to N do
If (j⬍⬎k) then A[k,j] :⫽ A[k,j] / A[k,k];
A[k,k] :⫽ 1.0 / A[k,k];
For i :⫽ 1 to N do
If (i⬍⬎k) then Begin
For j :⫽ 1 to N do
If (j⬍⬎k) then A[i,j] :⫽ A[i,j] ⫺ A[i,k]*A[k,j];
A[i,k] :⫽ ⫺A[i,k]*A[k,k];
End; {If i⬍⬎k}
End; {for k}
542
SOLUTION OF EQUATIONS BY MATRIX METHODS
A3
I3 ⫽ A⫺1
1 0 0 兩 12/24 ⫺1/4 ⫺5/24
0 1 0 兩 ⫺5/12
1/2
1/12
0 0 1 兩
1/24 ⫺1/4
7/24
冤
冥
Notice that through of these nine elementary row transformations, the original A matrix is transformed into the identity matrix and the original identity
matrix is transformed into A⫺1. Also note that A⫺1 obtained by this method
agrees exactly with the inverse obtained by the method of adjoints. This is
because any nonsingular matrix has a unique inverse.
It should be obvious that the quantity of work involved in inverting matrices increases greatly with the matrix size, since the number of necessary
row transformations is equal to the square of the number of rows or columns.
Because of this, it is not considered practical to invert large matrices by hand.
This work is done more conveniently with a computer. Since the procedure
of elementary row transformations is systematic, it is easily programmed.
Table B.1 shows algorithms, written in the BASIC, C, FORTRAN, and
Pascal programming languages, for calculating the inverse of any n ⫻ n nonsingular matrix A. Students should review the code in their preferred language
to gain familiarity with computer procedures.
B.6
EXAMPLE PROBLEM
Example B.2 Suppose that an EDM instrument is placed at point A in Figure
B.2 and a reflector is placed successively at B, C, and D. The observed values
AB, AC, and AD are shown in the figure. Calculate the unknowns X1, X2, and
X3 by matrix methods. The values observed are
AB ⫽ 125.27
AC ⫽ 259.60
AD ⫽ 395.85
Figure B.2 Observation of a line.
PROBLEMS
543
SOLUTION Formulate the basic equations:
1X1 ⫹ 0X2 ⫹ 0X3 ⫽ 125.27
1X1 ⫹ 1X2 ⫹ 0X3 ⫽ 259.60
1X1 ⫹ 1X2 ⫹ 1X3 ⫽ 395.85
Represented in matrix notation, these equations are AX ⫽ L. In this matrix
equation, the individual matrices are
冤
冥
1 0 0
A⫽ 1 1 0
1 1 1
冤冥
x1
X ⫽ x2
x3
冤 冥
125.27
L ⫽ 259.60
395.85
The solution in matrix notation is X ⫽ A⫺1L. Using elementary row transformation, the inverse of A is
冤
冥 冤
1 0 0 兩 1 0 0
1 0 0 兩
1
1 1 0 兩 0 1 0 → 0 1 0 兩 ⫺1
1 1 1 兩 0 0 1
0 0 1 兩
0
Solving X ⫽ A⫺1L, the unknowns are
冤
冥冤 冥 冤 冥
1
0 0
X ⫽ A⫺1L ⫽ ⫺1
1 0
0 ⫺1 1
125.27
125.27
259.60 ⫽ 134.33
395.85
136.25
X1 ⫽ 125.27
X2 ⫽ 134.33
X3 ⫽ 136.25
PROBLEMS
B.1
Describe when a 2 ⫻ 2 matrix has no inverse.
B.2
Find the inverse of A using the method of adjoints.
冤
冥
3 ⫺1 ⫺1
A ⫽ ⫺1
3 ⫺1
⫺1 ⫺1
3
冥
0 0
1 0
⫺1 1
544
SOLUTION OF EQUATIONS BY MATRIX METHODS
B.3
Find the inverse of A in Problem B.2 using elementary row transformations.
B.4
Solve the following system of linear equations using matrix methods.
x ⫹ 5y ⫽ ⫺8
⫺x ⫺ 2y ⫽ ⫺1
B.5
Solve the following system of linear equations using matrix methods.
x ⫹ y ⫺ z ⫽ ⫺8
3x ⫺ y ⫹ z ⫽ ⫺4
⫺x ⫹ 2y ⫹ 2z ⫽ 21
B.6
Compute the inverses of the following matrices.
A⫽
B.7
冋 册
8
5
3 12
冤
0
冥
⫺1
3
16
⫺8
2
3
冤
4 3
B ⫽ ⫺1 0
2 8
冥
7
4
10
Compute the inverses of the following matrices.
冤
冥
13 ⫺6
0
A ⫽ ⫺6 18 ⫺6
0 ⫺6 16
B.9
冋 册
Compute the inverses of the following matrices.
3 ⫺1
A ⫽ ⫺1
3
0 ⫺1
B.8
B⫽
冤
冥
1
2
6
B ⫽ 2 ⫺3
4
0
6 ⫺12
Solve the following matrix system.
冤
13 ⫺6
18
0 ⫺6
⫺6
0
冥冤 冥 冤
⫺6
16
Use the MATRIX software to do each problem.
B.10 Problem B.4
B.11 Problem B.5
B.12 Problem B.6
冥
A
740.02
B ⫽ 612.72
C
1072.22
PROBLEMS
545
B.13 Problem B.7
B.14 Problem B.8
B.15 Problem B.9
Programming Problems
B.16 Select one of the coded matrix inverse routines from Table B.1, enter
the code into a computer, and use it to solve Problem B.7. (Hint: Place
the code in Table B.1 in a separate subroutine/function/procedure to
be called from the main program.)
B.17 Add a block of code to the inverse routine in the language of your
choice that will inform the user when a matrix is singular.
B.18 Write a program that reads and writes a file with a nonsingular matrix;
finds its inverse, and write the results. Use this program to solve Problem B.7. (Hint: Place the reading, writing, and inversing code in
separate subroutines/functions/procedures to be called from the main
program. Provide a way to identify each matrix in the output file.)
B.19 Write a program that reads and writes a file containing a system of
equations written in matrix form, solves the system using matrix operations, and writes the solution. Use this program to solve Problem
B.9. (Hint: Place the reading, writing, and inversing code in separate
subroutines/functions/procedures to be called from the main program.
Provide a way to identify each matrix in the output file.)
APPENDIX C
NONLINEAR EQUATIONS AND
TAYLOR’S THEOREM
C.1
INTRODUCTION
In adjustment computations it is frequently necessary to deal with nonlinear
equations. For example, some observation equations relate observed quantities
to unknown parameters through the transcendental functions of sine, cosine,
or tangent; others relate them through terms raised to second- and higherorder powers. The task of solving a system of nonlinear equations is formidable. To facilitate the solution, a first-order Taylor series approximation may
be used to create a set of linear equations. The equations can then be solved
by matrix methods discussed in Appendix B.
C.2
TAYLOR SERIES LINEARIZATION OF NONLINEAR EQUATIONS
Suppose that the following equation relates a observed value L to its unknown
parameters x and y through nonlinear coefficients as
L ⫽ ƒ(x,y)
(C.1)
By Taylor’s theorem, the equation is represented as
L ⫽ ƒ(x,y)
⫽ ƒ(x0,y0) ⫹
⫹
546
(L/ x)0
(2L/ x2)0 2
(nL/ xn)0 n
dx ⫹
dx ⫹ 䡠 䡠 䡠 ⫹
dx
1!
2!
n!
(L/ y)0
(2L/ y2)0 2
(nL/ yn)0 n
dy ⫹
dy ⫹ 䡠 䡠 䡠 ⫹
dy ⫹ R
1!
2!
n!
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
(C.2)
C.3
NUMERICAL EXAMPLE
547
In Equation (C.2), x0 and y0 are approximations of x and y; ƒ(x0,y0) is the
nonlinear function evaluated at these approximations; R is the remainder, and
dx and dy are corrections to the initial approximations, such that
x ⫽ x0 ⫹ dx
(C.3)
y ⫽ y0 ⫹ dy
A more exact Taylor series approximation is obtained by increasing the
value of n in Equation (C.2). However, as the order of each successive term
increases, its significance in the overall expression decreases. If all terms
containing derivatives higher than the first are dropped, the following linear
expression is obtained:
L ⫽ ƒ(x,y) ⫽ ƒ(x0,y0) ⫹
冉冊
L
x
0
dx ⫹
冉冊
L
y
dy
(C.4)
0
Once the initial approximations are selected, the only unknowns in Equation (C.4) are the corrections dx and dy. Of course, by dropping the
higher-order terms from the Taylor series, Equation (C.4) becomes only a
good approximation of the original equation. However, in the solution, an
iterative procedure can be followed that yields accurate answers. This iterative
procedure uses the following steps:
Step 1: Determine initial approximations for the unknowns. They may be
obtained by guessing or from observations. It should be understood that
the closer the initial approximations are to the final solution, the faster it
will be obtained. For some problems, initial approximations can be obtained from graphical solutions or computed from available data or observations. For others, the determination of initial approximations can involve
considerable computational effort.
Step 2: Substitute the initial approximations into Equation (C.4) and solve
for the corrections dx and dy.
Step 3: Calculate revised values of x and y using Equations (C.3).
Step 4: Using these newly revised values for x and y, repeat steps 2 and 3.
Step 5: Continue the procedure until the corrections dx and dy are small
enough to bring x and y within tolerable accuracy. When this occurs, the
solution is said to have converged.
C.3
NUMERICAL EXAMPLE
To clarify this procedure further, a numerical example will be solved.
548
NONLINEAR EQUATIONS AND TAYLOR’S THEOREM
Example C.1 Linearize the following pair of nonlinear equations containing
the two unknowns x and y.
F(x,y) ⫽ x ⫹ y ⫺ 2y2 ⫽ ⫺4
G(x,y) ⫽ x2 ⫹ y2 ⫽ 8
SOLUTION Determine the partial derivative for each equation with respect
to each unknown.
F
⫽1
x
F
⫽ 1 ⫺ 4y
y
G
⫽ 2x
x
F
⫽ 2y
y
Compute initial approximations for each unknown. An estimate of x ⫽ 1
and y ⫽ 1 is used for the approximations initially.
First iteration: Write the linearized equations in the form of Equation
(C.4).
F(x,y) ⫽ 1 ⫹ 1 ⫺ 2(1)2 ⫹ dx ⫹ [1 ⫺ 4(1)] dy ⫽ ⫺4
G(x,y) ⫽ (1)2 ⫹ (1)2 ⫹ 2(1) dx ⫹ 2(1) dy ⫽ 8
From the two equations above, solve for the unknowns dx and dy according
to Equation (C.3):
dx ⫽ 1.25
and dy ⫽ 1.75
Using this solution, determine updated values for x and y:
x ⫽ x0 ⫹ dx ⫽ 1.00 ⫹ 1.25 ⫽ 2.25
y ⫽ y0 ⫹ dy ⫽ 1.00 ⫹ 1.75 ⫽ 2.75
Second iteration: Continue the procedure demonstrated for the first
iteration.
F ⫽ 2.25 ⫹ 2.75 ⫺ 2(2.75)2 ⫹ dx ⫹ [1 ⫺ 4(2.75)] dy ⫽ ⫺4
G ⫽ (2.25)2 ⫹ (2.75)2 ⫹ 2(2.25) dx ⫹ 2(2.75) dy ⫽ 8
From the two equations above, dx ⫽ ⫺0.25 and dy ⫽ ⫺0.64, from which
C.4
USING MATRICES TO SOLVE NONLINEAR EQUATIONS
549
x ⫽ x0 ⫹ dx ⫽ 2.25 ⫺ 0.25 ⫽ 2.00
y ⫽ y0 ⫹ dy ⫽ 2.75 ⫺ 0.64 ⫽ 2.11
Third iteration:
F ⫽ 2 ⫹ 2.11 ⫺ 2(2.11)2 ⫹ dx ⫹ [1 ⫺ 4(2.11)] dy ⫽ ⫺4
G ⫽ (2)2 ⫹ (2.11)2 ⫹ 2(2) dx ⫹ 2(2.11) dy ⫽ 8
From the two equations above, dx ⫽ 0.00 and dy ⫽ ⫺0.11, from which
x ⫽ x0 ⫹ dx ⫽ 2.00 ⫹ 0.00 ⫽ 2.00
y ⫽ y0 ⫹ dy ⫽ 2.11 ⫺ 0.11 ⫽ 2.00
Fourth iteration:
F ⫽ 2 ⫹ 2 ⫺ 2(2)2 ⫹ dx ⫹ [1 ⫺ 4(2)] dy ⫽ ⫺4
G ⫽ (2)2 ⫹ (2)2 ⫹ 2(2)dx ⫹ 2(2) dy ⫽ 8
Using the two equations above, the corrections to x and y are zero to the
nearest hundredth. Thus, the solution has converged and the values of x ⫽
2.00 and y ⫽ 2.00 are the desired unknowns. Note that the initial values for
the initial approximations were relatively poor and four iterations were required to find the final solution. However, had better estimates been made
(say, x0 ⫽ 2.1 and y0 ⫽ 1.9), the solution would have converged in one or
two iterations and saved computational effort. Fortunately, there are accepted
computational procedures to determine close approximations in many surveying problems.
C.4
USING MATRICES TO SOLVE NONLINEAR EQUATIONS
The example of Section C.3 could be solved using matrix methods. However,
as in the algebraic approach, the equations must be linearized using Taylor’s
series. To facilitate linearization using Taylor’s theorem, the Jacobian matrix
(a matrix consisting of the partial derivatives taken with respect to the unknown variables) is formed. This is the coefficient matrix of the linearized
equations. The Jacobian matrix for the example of Section C.3 is
550
NONLINEAR EQUATIONS AND TAYLOR’S THEOREM
J⫽
冤 冥
F
x
F
y
G
x
G
y
In the Jacobian matrix above, the first column contains the partial derivative
for each equation with respect to x, and the second column contains the partial
derivative of each equation with respect to y. The linearized form of the
equations may then be expressed in matrix notation as
(C.5)
JX ⫽ K
In Equation (C.5), J is the Jacobian matrix, X the matrix of unknown
corrections dx and dy, and K the matrix of constants. Specifically, for the
example of Section C.3, these matrices are
J⫽
1
2x0
1 ⫺ 4y0
2y0
X⫽
dx
dy
K⫽
⫺4 ⫺ F(x0, y0)
8 ⫺ G(x0, y0)
where F(x0,y0) and G(x0,y0) are the equations F and G solved at the initial
approximations of x0 and y0.
Beginning with a set of initial approximations x0 and y0, the J and K
matrices of Equation (C.5) are formed. X is computed using the matrix methods presented in Appendix B. Having updated the unknowns according to
Equations (C.3), the J and K matrices are formed again and the solution for
X is computed. This procedure is iterated until convergence is achieved.
C.5
SIMPLE MATRIX EXAMPLE
Example C.2 Find the solution of the nonlinear system of equations shown
below using matrix methods.
F(x, y) ⫽ x2 ⫹ 3xy ⫺ 4y2 ⫽ 6
G(x, y) ⫽ x ⫹ xy ⫺ y2 ⫽ 3
SOLUTION The partial derivatives of functions F and G with respect to
the unknowns, x and y, are
F
⫽ 2x ⫹ 3y
x
F
⫽ 3x ⫺ 8y
y
G
⫽1⫹y
x
G
⫽ x ⫺ 2y
y
Thus, the Jacobian matrix is
C.6
J⫽
冤 冥
F
x
F
y
G
x
G
y
⫽
PRACTICAL EXAMPLE
551
2x0 ⫹ 3y0 3x0 ⫺ 8y0
1 ⫹ y0
x0 ⫺ 2y0
and the system of equations to solve is
2x0 ⫹ 3y0 3x0 ⫺ 8y0
1 ⫹ y0
x0 ⫺ 2y0
dx
6 ⫺ F(x0, y0)
⫽
dy
3 ⫺ G(x0, y0)
First iteration: Using x0 ⫽ 3 and y0 ⫽ 0 yields
6 9
1 3
dx
6⫺9
⫺3
⫽
⫽
dy
3⫺3
0
The determinant for the Jacobian matrix above is 3(6) ⫺ 9(1) ⫽ 9, and thus
the matrix solution is
1
dx
3 9
⫽
dy
9 ⫺1 6
⫺3
0
⫽
⫺1.0
0.3
Applying Equation (C.3), initial approximations for a second iteration are
⫺1
2.0
x
x
dx
3
⫽ 0 ⫹
⫽
⫹
⫽
y
y0
dy
0
0.3
0.3
Second iteration:
4.9 3.6
1.3 1.4
dx
6 ⫺ 5.44
0.56
⫽
⫽
dy
3 ⫺ 2.51
0.49
From this, dx and dy are found to be 0.45 and 0.77, respectively. This
makes the approximations of x and y for the third iteration 1.55 and 1.07,
respectively. The procedures are followed until the final solution for x and y
is found to be 2.00 and 1.00, respectively. Again, fewer iterations would have
been required if the initial approximations had been closer to the final values.
C.6
PRACTICAL EXAMPLE
Example C.3 Assume that the x and y coordinates of three points on a circle
have been observed. Their coordinates are (9.4, 5.6), (7.6, 7.2), and (3.8, 4.8),
respectively. The equation for a circle with center (h,k) and radius r is
552
NONLINEAR EQUATIONS AND TAYLOR’S THEOREM
(x ⫺ h)2 ⫹ (y ⫺ k)2 ⫽ r 2. Determine the coordinates of the center of the
circle and its radius.
SOLUTION The equation of a circle is rewritten as C(h,k,r) ⫽ (x ⫺ h)2 ⫹
(y ⫺ k)2 ⫺ r 2 ⫽ 0. The partial derivatives with respect to the unknowns h,
k, and r are
C
⫽ ⫺2(x ⫺ h)
h
C
⫽ ⫺2(y ⫺ k)
k
C
⫽ ⫺2r
r
For each point observed, one equation is written, resulting in a system of
three equations and three unknowns. The general linearized form of these
equations expressed using matrices is
冤 冥
C1
h
C1
k
C1
r
C2
h
C2
k
C2
r
C3
h
C2
k
C3
r
冤冥 冤
冥
dh
0 ⫺ [(x1 ⫺ h0)2 ⫹ (y1 ⫺ k0)2 ⫺ r 02]
dk ⫽ 0 ⫺ [(x2 ⫺ h0)2 ⫹ (y2 ⫺ k0)2 ⫺ r 20]
dr
0 ⫺ [(x3 ⫺ h0)2 ⫹ (y3 ⫺ k0)2 ⫺ r 02]
(C.6)
After taking partial derivatives, Equation (C.6) becomes
冤
⫺2(x1 ⫺ h0)
⫺2(x2 ⫺ h0)
⫺2(x3 ⫺ h0)
⫺2(y1 ⫺ k0)
⫺2(y2 ⫺ k0)
⫺2(y3 ⫺ k0)
冤
冥冤 冥
冥
⫺2r0
⫺2r0
⫺2r0
dh
dk
dr
0 ⫺ [(x1 ⫺ h0)2 ⫹ (y1 ⫺ k0)2 ⫺ r 02]
⫽ 0 ⫺ [(x2 ⫺ h0)2 ⫹ (y2 ⫺ k0)2 ⫺ r 20]
0 ⫺ [(x3 ⫺ h0)2 ⫹ (y3 ⫺ k0)2 ⫺ r 02]
(C.7)
Equations (C.7) can be simplified by multiplying each side by ⫺1/2. The
resulting equations are
冤
冥冤 冥 冤
(x1 ⫺ h0) (y1 ⫺ k0) r0
(x2 ⫺ h0) (y2 ⫺ k0) r0
(x3 ⫺ h0) (y3 ⫺ k0) r0
冥
dh
0.5[(x1 ⫺ h0)2 ⫹ (y1 ⫺ k0)2 ⫺ r 02]
dk ⫽ 0.5[(x2 ⫺ h0)2 ⫹ (y2 ⫺ k0)2 ⫺ r 20]
dr
0.5[(x3 ⫺ h0)2 ⫹ (y3 ⫺ k0)2 ⫺ r 02]
(C.8)
Assuming approximate initial values for h, k, and r as 7, 4.5, and 3, respectively, Equations (C.8) are
C.6
冤
9.4 ⫺ 7
7.6 ⫺ 7
3.8 ⫺ 7
5.6 ⫺ 4.5
7.2 ⫺ 4.5
4.8 ⫺ 4.5
PRACTICAL EXAMPLE
冥冤 冥 冤
3
3
3
dh
dk
dr
553
冥
0.5[(9.4 ⫺ 7)2 ⫹ (5.6 ⫺ 4.5)2 ⫺ 32]
⫽ 0.5[(7.6 ⫺ 7)2 ⫹ (7.2 ⫺ 4.5)2 ⫺ 32]
0.5[(3.8 ⫺ 7)2 ⫹ (4.8 ⫺ 4.5)2 ⫺ 32]
(C.9)
Simplifying Equations (C.9) yields
冤
冥冤 冥 冤
2.4 1.1 3
0.6 2.7 3
⫺3.2 0.3 3
冥
dh
⫺1.015
dk ⫽ ⫺0.675
dr
0.665
Solving this system gives results of
冤冥 冤
冥
⫺0.28462
dh
dk ⫽ ⫺0.10769
dr
⫺0.07115
After applying these changes to the initial approximations, updated values
for h, k, and r of 6.7154, 4.3923, and 2.9288, respectively, are obtained. The
second iteration results in corrections of 0, 0, and 0.014945. Since the correction for r is still comparatively large, the iteration process must be continued. After the third iteration, suitable convergence was achieved. The final
values for h, k, and r are 6.72, 4.39, and 2.94, respectively, which are within
0.00001 of a perfect solution.
Sometimes, more than one method is available for solving a problem. For
example, in Example C.3 an alternative linear form of the equation of a circle
could have been used. That equation is x2 ⫹ y2 ⫹ 2dx ⫹ 2ey ⫹ ƒ ⫽ 0, where
the center of the circle is at (⫺d, ⫺e) and the circle’s radius is
d 2 ⫹ e2 ⫺ ƒ. Note that the equation is linear in terms of its unknowns (d,
e, ƒ) and thus iterations are not necessary in solving for the unknowns. Writing a rearranged form of this linear equation for each of three measured sets
of (x, y) coordinates yields
2dx1 ⫹ 2ey1 ⫹ ƒ ⫽ ⫺(x12 ⫹ y12)
2dx2 ⫹ 2ey2 ⫹ ƒ ⫽ ⫺(x22 ⫹ y22)
2dx3 ⫹ 2ey3 ⫹ ƒ ⫽ ⫺(x32 ⫹ y32)
Equations (C.10) can in turn be represented in matrix notation as
(C.10)
554
NONLINEAR EQUATIONS AND TAYLOR’S THEOREM
冤
2x1 2y1 1
2x2 2y2 1
2x3 2y3 1
冥冤 冥 冤
冥
d
⫺(x12 ⫹ y12)
e ⫽ ⫺(x22 ⫹ y22)
ƒ
⫺(x32 ⫹ y32)
Solving this matrix system, the center of the circle is again found to be
(6.72, 4.39), and its radius is determined to be 2.94.
In this appendix, the Taylor series has been applied to solve for the unknowns in nonlinear equations. Many equations in surveying, geodesy, and
photogrammetry are nonlinear. In surveying, examples include the distance
and angle formulas, which are nonlinear in terms of station coordinates. The
Taylor series is used to linearize these equations and find least squares solutions. Thus, when performing least squares adjustments of plane observations,
the techniques presented in this appendix must be used in the solutions.
PROBLEMS
C.1
Solve for the unknowns x and y in the following nonlinear equations
using Taylor’s theorem. (Use x0 ⫽ 5 and y0 ⫽ 5 for initial approximations.)
x2y ⫺ 3x2 ⫽ 75
x2 ⫺ y ⫽ 19
C.2
Solve for the unknown values of x, y, and z in the following three
nonlinear equations using the Taylor series. (Use x0 ⫽ y0 ⫽ z0 ⫽ 2 for
initial approximations.)
x2 ⫺ y2 ⫹ 2xy ⫹ z ⫽ 4
⫺x ⫹ y ⫹ z ⫽ 4
⫺2x2 ⫺ y ⫹ z3 ⫽ 23
C.3
Use the MATRIX software to solve Problem C.1.
C.4
Use the MATRIX software to solve Problem C.2.
C.5
Find the center and radius of a circle using the equation (x ⫺ h)2 ⫹
(y ⫺ k)2 ⫽ r 2 given the coordinates of points A, B, and C on the circle.
Follow the procedures discussed in Section C.5. Use initial approximations of h0 ⫽ 5, k0 ⫽ 4, and r0 ⫽ 2 for the first iteration.
A: (7.2,5.2)
C.6
B: (4.0, 6.4)
C: (4.0, 2.4)
Repeat Problem C.5 using the linear equation: x2 ⫹ y2 ⫹ 2dx ⫹ 2ey
⫹ ƒ ⫽ 0.
PROBLEMS
555
C.7
Repeat Problem C.5 using the points A: (0.50, ⫺0.70), B: (1.00, 0.00),
C: (0.70, 0.70). Use initial approximations of h0 ⫽ 0, k0 ⫽ 0, and r0
⫽ 1.
C.8
Repeat Problem C.7 using the linear equation x2 ⫹ y2 ⫹ 2dx ⫹ 2ey
⫹ ƒ ⫽ 0.
C.9
Use the ADJUST software to solve Problem C.5.
C.10
Use the ADJUST software to solve Problem C.7.
C.11
The distance formula between two stations i and j is Dij ⫽
(xj ⫺ xi)2 ⫹ (yj ⫺ yi)2. Write the linearized form of in terms of the
variables xi, yi, xj, and yj.
C.12
The azimuth formula between two stations i and j is ␣ij ⫽ tan⫺1
[(xj ⫺ xi)/(yj ⫺ yi)]. Write the linearized form of this equation in terms
of the variables xi, yi, xj, and yj.
C.13
The formula for an angle ∠jik is ␣ik ⫺ ␣ij, where ␣ is defined in
Problem C.12. Write the linearized form of this equation in terms of
the variables xi, yi, xj, yj, xk, and yk.
Programming Problems
C.14
Create a programmed package that solves Problem C.6.
C.15
Create a programmed package that solves Problem C.5.
APPENDIX D
NORMAL ERROR DISTRIBUTION
CURVE AND OTHER
STATISTICAL TABLES
D.1 DEVELOPMENT OF THE NORMAL DISTRIBUTION
CURVE EQUATION
In Section 2.4 the histogram and frequency polygon were presented as methods for graphical portrayal of random error distributions. If a large number
of these distributions were examined for sets, measurements in surveying,
geodesy, and photogrammetry, it would be found that they conform to normal
(or Gaussian) distributions. The general laws governing normal distributions
are stated as follows:
1. Positive and negative errors occur with equal probability and equal
frequency.
2. Small errors are more common than large errors.
3. Large errors seldom occur, and there is a limit to the size of the greatest
random error that will occur in any set of observations.
A curve that conforms to these laws, plotted with the size of the error on
the abscissa and the probability of occurrence on the ordinate, is shown in
Figure 3.3. This curve is repeated in Figure D.1 and is called the normal
distribution curve, the normal curve of error, or simply the probability curve.
A smooth curve of this same shape would be obtained if for a very large
group of measurements, a histogram were plotted with an infinitesimally small
class interval. In this section, the equation for this curve is developed.
Assume that the normal distribution curve is continuous and that the probability of an error occurring between x and x ⫹ dx is given by the function
556
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
D.1
DEVELOPMENT OF THE NORMAL DISTRIBUTION CURVE EQUATION
557
Figure D.1 Normal distribution curve.
y ⫽ ƒ(x). Further assume that this is the equation for the probability curve.
The form of ƒ(x) will now be determined. Since as explained in Chapter 3,
probabilities are equivalent to areas under the probability curve, the probabilities of errors occurring within the ranges (x1 and x1 ⫹ dx1), (x2 and x2 ⫹
dx2), and so on, are ƒ(x1) dx1, ƒ(x2) dx2, . . . , ƒ(xn) dxn. The total area under
the probability curve represents the total probability or simply 1. Then for a
finite number of possible errors,
ƒ(x1) dx1 ⫹ ƒ(x2) dx2 ⫹ 䡠 䡠 䡠 ⫹ ƒ(xn) dxn ⫽ 1
(D.1)
If the total range of errors x1, x2, . . . , xn is between plus and minus 1,
considering an infinite number of errors that makes the curve continuous, the
area under the curve can be set equal to
冕
1
1⫽
ƒ(x) dx
⫺1
But because the area under the curve from ⫹l to ⫹⬁ and from ⫺l to ⫺⬁ is
essentially zero, the integration limits are extended to Ⳳ⬁, as
冕
⬁
1⫽
⫺⬁
冕
⬁
ƒ(x) dx ⫽
y dx
(D.2)
⫺⬁
Now suppose that the quantity M has been measured and that it is equal
to some function of n unknown parameters z1, z2, . . . , zn such that M ⫽ ƒ(z1,
z2, . . . , zn). Also let x1, x2, . . . , xm be the errors of m observations M1, M2,
. . . , Mm, and let ƒ(x1) dx1, ƒ(x2) dx2, . . . , ƒ(xm) dxm be the probabilities of
errors falling within the ranges (x1 and dx1), (x2 and dx2), and so on. By
558
NORMAL ERROR DISTRIBUTION CURVE AND OTHER STATISTICAL TABLES
Equation (3.1), the probability P of the simultaneous occurrence of all of
these errors is equal to the product of the individual probabilities; thus,
P ⫽ [ƒ(x1) dx1] [ƒ(x2) dx2] 䡠 䡠 䡠 [ƒ(xm) dxm]
Then, by logs,
log P ⫽ log ƒ(x1) ⫹ log ƒ(x2) ⫹ 䡠 䡠 䡠 ⫹ log ƒ(xm)
⫹ log dx1 ⫹ log dx2 ⫹ 䡠 䡠 䡠 ⫹ log dxm
(D.3)
The most probable values of the errors will occur when P is maximized
or when the log of P is maximized. To maximize a function, it is differentiated
with respect to each unknown parameter z, and the results set equal to zero.
After logarithmic differentiation of Equation (D.3), the following n equations
result (note that the dx’s are constants independent of the z’s, and therefore
their differentials with respect to the z’s are zero):
1 dƒ(x1) dx1
1 dƒ(x2) dx2
1 dƒ(xm) dxm
1 P
⫽
⫹
⫹䡠䡠䡠⫹
⫽0
P z1 ƒ(x1) dx1 dz1 ƒ(x2) dx2 dz1
ƒ(xm) dxm dz1
1 P
1 dƒ(x1) dx1
1 dƒ(x2) dx2
1 dƒ(xm) dxm
⫽
⫹
⫹䡠䡠䡠⫹
⫽0
P z2 ƒ(x1) dx1 dz2 ƒ(x2) dx2 dz2
ƒ(xm) dxm dz2
⯗
1 P
1 dƒ(x1) dx1
1 dƒ(x2) dx2
1 dƒ(xm) dxm
⫽
⫹
⫹䡠䡠䡠⫹
⫽0
P zn ƒ(x1) dx1 dzn ƒ(x2) dx2 dzn
ƒ(xm) dxm dzn
(D.4)
Now let
ƒ(x) ⫽
dƒ(x)
dx
(D.5)
Substituting Equation (D.5) into (D.4) gives
ƒ(xm) dxm
ƒ(x1) dx1 ƒ(x2) dx2
⫹
⫹䡠䡠䡠⫹
⫽0
ƒ(x1) dz1
ƒ(x2) dz1
ƒ(xm) dz1
ƒ(x1) dx1 ƒ(x2) dx2
ƒ(xm) dxm
⫹
⫹䡠䡠䡠⫹
⫽0
ƒ(x1) dz2
ƒ(x2) dz2
ƒ(xm) dz2
⯗
ƒ(x1) dx1 ƒ(x2) dx2
ƒ(xm) dxm
⫹
⫹䡠䡠䡠⫹
⫽0
ƒ(x1) dzn
ƒ(x2) dzn
ƒ(xm) dzn
(D.6)
D.1
DEVELOPMENT OF THE NORMAL DISTRIBUTION CURVE EQUATION
559
Thus far, ƒ(x) and ƒ(x) are general, regardless of the number of unknown
parameters. Now consider the special case where there is only one unknown
z and M1, M2, . . . , Mm are m observed values of z. If z* is the true value of
the quantity, the errors associated with the observations are
x1 ⫽ z* ⫺ M1, x2 ⫽ z* ⫺ M2, . . . , xm ⫽ z* ⫺ Mm
(D.7)
Differentiating Equation (D.7) gives
1⫽
dx1 dx2
dx
⫽
⫽䡠䡠䡠⫽ m
dz
dz
dz
(D.8)
Then for this special case, substituting Equations (D.7) and (D.8) into Equations (D.6), they reduce to a single equation:
ƒ(z* ⫺ Mm)
ƒ(z* ⫺ M1) ƒ(z* ⫺ M2)
⫹
⫹䡠䡠䡠⫹
⫽0
ƒ(z* ⫺ M1)
ƒ(z* ⫺ M2)
ƒ(z* ⫺ Mm)
(D.9)
Equation (D.9) for this special case in consideration is also general for any
value of m and for any observed values M1, M2, . . . , Mm. Thus, let the values
of M be
M2 ⫽ M3 ⫽ 䡠 䡠 䡠 ⫽ Mm ⫽ M1 ⫺ mN
where N is chosen for convenience as N ⫽ (M1 ⫺ M2)/m.
The arithmetic mean is the most probable value for this case of a single
quantity having been observed several times; therefore, z*, the most probable
value in this case, is
z* ⫽
⫽
M1 ⫹ M2 ⫹ 䡠 䡠 䡠 ⫹ Mn
m
M1 ⫹ (m ⫺ 1)(M1 ⫺ mN)
m
⫽ M1 ⫺ mN ⫹ N
⫽ M1 ⫺ N(m ⫺ 1)
z* ⫺ M1 ⫽ ⫺N(m ⫺ 1) ⫽ N(1 ⫺ m)
(D.10)
Recall that N ⫽ (M1 ⫺ M2)/m, from which M1 ⫽ mN ⫹ M2. Substituting into
Equation (D.10) gives
z* ⫺ (mN ⫹ M2) ⫽ N(1 ⫺ m)
z* ⫺ M2 ⫽ N
560
NORMAL ERROR DISTRIBUTION CURVE AND OTHER STATISTICAL TABLES
Similarly, since N ⫽ (M1 ⫺ M3)/m ⫽ (M1 ⫺ M4)/m, and so on,
z* ⫺ M3 ⫽ N
z* ⫺ M4 ⫽ N
⯗
Substituting these expressions into Equation (D.9) yields
ƒ[n(1 ⫺ m)] (m ⫺ 1)ƒ(N)
⫹
⫽0
ƒ[N(1 ⫺ M)]
ƒ(N)
(D.11)
Rearranging yields
ƒ(N)
ƒ[N(1 ⫺ m)]
⫽
⫽ constant
Nƒ[N(1 ⫺ m)](1 ⫺ m) ƒ(N)N
because N in this case is a constant. Thus,
ƒ(x)
⫽ constant ⫽ K
xƒ(x)
(D.12)
Substituting Equation (D.5) into Equation (D.12) yields
dƒ(x)
dx
ƒ(x) ⫽ xƒ(x)K ⫽
from which dƒ(x)/dx ⫽ xƒ(x)K. Integrating gives
loge ƒ(x) ⫽ –12 Kx2 ⫹ C1
2/2
ƒ(x) ⫽ eC1eKx
But letting
eC1 ⫽ C
then
2
ƒ(x) ⫽ Ce(Kx ) / 2
(D.13)
In Equation (D.13), since ƒ(x) decreases as x increases, the exponent must
be negative. Arbitrarily letting
h⫽
冪K2
(D.14)
D.1
DEVELOPMENT OF THE NORMAL DISTRIBUTION CURVE EQUATION
561
and incorporating the negative into Equation (D.13), there results
2 2
ƒ(x) ⫽ Ce⫺h x
(D.15)
To find the value of the constant C, substitute Equation (D.15) into Equation
(D.2):
冕
⬁
⫺⬁
2 2
Ce⫺h x dx ⫽ 1
Also, arbitrarily set t ⫽ hx; then dt ⫽ h dx and dx ⫽ dt/h, from which, after
changing variables, we obtain
C
⫽
h
冕
⬁
2
⫺⬁
e⫺t dt ⫽ 1
The value of the definite integral is 兹, from which1
C
h兹
C⫽
⫽1
h
(D.16)
兹
Substituting Equation (D.16) into Equation (D.15) gives
ƒ(x) ⫽
h
兹
2 2
e⫺h x
(D.17)
Note that from Equation (D.14) that h ⫽ 兹K/2. For the normal distribution,
K ⫽ 1/ 2. Substituting this into Equation (D.17) gives
ƒ(x) ⫽
1
1
2 2
2
2
e⫺(1 / 2 )x ⫽
e⫺x / 2
2
兹2
兹2
(D.18)
where the terms are as defined for Equation (3.2).
This is the general equation for the probability curve, having been derived
in this instance from the consideration of a special case. In Table D.1, values
for areas under the standard normal distribution function from negative infinity to t are tabulated.
1
The technique of integrating this nonelementary function is beyond the scope of this book but
can be found in advanced references.
562
NORMAL ERROR DISTRIBUTION CURVE AND OTHER STATISTICAL TABLES
TABLE D.1 Percentage Points for the Standard Normal Distribution Function
Nz(t) ⫽
冕
⬁
1
⫺⬁
兹2
e⫺x2 / 2 dx
t
t
0
1
2
3
4
5
6
7
8
9
⫺3.2
⫺3.1
⫺3.0
⫺2.9
⫺2.8
⫺2.7
⫺2.6
⫺2.5
⫺2.4
⫺2.3
⫺2.2
⫺2.1
⫺2.0
⫺1.9
⫺1.8
⫺1.7
⫺1.6
⫺1.5
⫺1.4
⫺1.3
⫺1.2
⫺1.1
⫺1.0
⫺0.9
⫺0.8
⫺0.7
⫺0.6
⫺0.5
⫺0.4
⫺0.3
⫺0.2
⫺0.1
⫺0.0
0.00069
0.00097
0.00135
0.00187
0.00256
0.00347
0.00466
0.00621
0.00820
0.01072
0.01390
0.01786
0.02275
0.02872
0.03593
0.04457
0.05480
0.06681
0.08076
0.09680
0.11507
0.13567
0.15866
0.18406
0.21186
0.24196
0.27425
0.30854
0.34458
0.38209
0.42074
0.46017
0.50000
0.00066
0.00094
0.00131
0.00181
0.00248
0.00336
0.00453
0.00604
0.00798
0.01044
0.01355
0.01743
0.02222
0.02807
0.03515
0.04363
0.05370
0.06552
0.07927
0.09510
0.11314
0.13350
0.15625
0.18141
0.20897
0.23885
0.27093
0.30503
0.34090
0.37828
0.41683
0.45620
0.49601
0.00064
0.00090
0.00126
0.00175
0.00240
0.00326
0.00440
0.00587
0.00776
0.01017
0.01321
0.01700
0.02169
0.02743
0.03438
0.04272
0.05262
0.06426
0.07780
0.09342
0.11123
0.13136
0.15386
0.17879
0.20611
0.23576
0.26763
0.30153
0.33724
0.37448
0.41294
0.45224
0.49202
0.00062
0.00087
0.00122
0.00169
0.00233
0.00317
0.00427
0.00570
0.00755
0.00990
0.01287
0.01659
0.02118
0.02680
0.03362
0.04182
0.05155
0.06301
0.07636
0.09176
0.10935
0.12924
0.15151
0.17619
0.20327
0.23270
0.26435
0.29806
0.33360
0.37070
0.40905
0.44828
0.48803
0.00060
0.00084
0.00118
0.00164
0.00226
0.00307
0.00415
0.00554
0.00734
0.00964
0.01255
0.01618
0.02068
0.02619
0.03288
0.04093
0.05050
0.06178
0.07493
0.09012
0.10749
0.12714
0.14917
0.17361
0.20045
0.22965
0.26109
0.29460
0.32997
0.36693
0.40517
0.44433
0.48405
0.00058
0.00082
0.00114
0.00159
0.00219
0.00298
0.00402
0.00539
0.00714
0.00939
0.01222
0.01578
0.02018
0.02559
0.03216
0.04006
0.04947
0.06057
0.07353
0.08851
0.10565
0.12507
0.14686
0.17106
0.19766
0.22663
0.25785
0.29116
0.32636
0.36317
0.40129
0.44038
0.48006
0.00056
0.00079
0.00111
0.00154
0.00212
0.00289
0.00391
0.00523
0.00695
0.00914
0.01191
0.01539
0.01970
0.02500
0.03144
0.03920
0.04846
0.05938
0.07215
0.08691
0.10383
0.12302
0.14457
0.16853
0.19489
0.22363
0.25463
0.28774
0.32276
0.35942
0.39743
0.43644
0.47608
0.00054
0.00076
0.00107
0.00149
0.00205
0.00280
0.00379
0.00508
0.00676
0.00889
0.01160
0.01500
0.01923
0.02442
0.03074
0.03836
0.04746
0.05821
0.07078
0.08534
0.10204
0.12100
0.14231
0.16602
0.19215
0.22065
0.25143
0.28434
0.31918
0.35569
0.39358
0.43251
0.47210
0.00052
0.00074
0.00104
0.00144
0.00199
0.00272
0.00368
0.00494
0.00657
0.00866
0.01130
0.01463
0.01876
0.02385
0.03005
0.03754
0.04648
0.05705
0.06944
0.08379
0.10027
0.11900
0.14007
0.16354
0.18943
0.21770
0.24825
0.28096
0.31561
0.35197
0.38974
0.42858
0.46812
0.00050
0.00071
0.00100
0.00139
0.00193
0.00264
0.00357
0.00480
0.00639
0.00842
0.01101
0.01426
0.01831
0.02330
0.02938
0.03673
0.04551
0.05592
0.06811
0.08226
0.09853
0.11702
0.13786
0.16109
0.18673
0.21476
0.24510
0.27760
0.31207
0.34827
0.38591
0.42465
0.46414
D.1
DEVELOPMENT OF THE NORMAL DISTRIBUTION CURVE EQUATION
563
TABLE D.1 (Continued )
Nz(t) ⫽
冕
⬁
1
⫺⬁
兹2
e⫺x2 / 2 dx
t
t
0.0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
2.0
2.1
2.2
2.3
2.4
2.5
2.6
2.7
2.8
2.9
3.0
3.1
3.2
0
1
2
3
4
5
6
7
8
9
0.50000
0.53983
0.57926
0.61791
0.65542
0.69146
0.72575
0.75804
0.78814
0.81594
0.84134
0.86433
0.88493
0.90320
0.91924
0.93319
0.94520
0.95543
0.96407
0.97128
0.97725
0.98214
0.98610
0.98928
0.99180
0.99379
0.99534
0.99653
0.99744
0.99813
0.99865
0.99903
0.99931
0.50399
0.54380
0.58317
0.62172
0.65910
0.69497
0.72907
0.76115
0.79103
0.81859
0.84375
0.86650
0.88686
0.90490
0.92073
0.93448
0.94630
0.95637
0.96485
0.97193
0.97778
0.98257
0.98645
0.98956
0.99202
0.99396
0.99547
0.99664
0.99752
0.99819
0.99869
0.99906
0.99934
0.50798
0.54776
0.58706
0.62552
0.66276
0.69847
0.73237
0.76424
0.79389
0.82121
0.84614
0.86864
0.88877
0.90658
0.92220
0.93574
0.94738
0.95728
0.96562
0.97257
0.97831
0.98300
0.98679
0.98983
0.99224
0.99413
0.99560
0.99674
0.99760
0.99825
0.99874
0.99910
0.99936
0.51197
0.55172
0.59095
0.62930
0.66640
0.70194
0.73565
0.76730
0.79673
0.82381
0.84849
0.87076
0.89065
0.90824
0.92364
0.93699
0.94845
0.95818
0.96638
0.97320
0.97882
0.98341
0.98713
0.99010
0.99245
0.99430
0.99573
0.99683
0.99767
0.99831
0.99878
0.99913
0.99938
0.51595
0.55567
0.59483
0.63307
0.67003
0.70540
0.73891
0.77035
0.79955
0.82639
0.85083
0.87286
0.89251
0.90988
0.92507
0.93822
0.94950
0.95907
0.96712
0.97381
0.97932
0.98382
0.98745
0.99036
0.99266
0.99446
0.99585
0.99693
0.99774
0.99836
0.99882
0.99916
0.99940
0.51994
0.55962
0.59871
0.63683
0.67364
0.70884
0.74215
0.77337
0.80234
0.82894
0.85314
0.87493
0.89435
0.91149
0.92647
0.93943
0.95053
0.95994
0.96784
0.97441
0.97982
0.98422
0.98778
0.99061
0.99286
0.99461
0.99598
0.99702
0.99781
0.99841
0.99886
0.99918
0.99942
0.52392
0.56356
0.60257
0.64058
0.67724
0.71226
0.74537
0.77637
0.80511
0.83147
0.85543
0.87698
0.89617
0.91309
0.92785
0.94062
0.95154
0.96080
0.96856
0.97500
0.98030
0.98461
0.98809
0.99086
0.99305
0.99477
0.99609
0.99711
0.99788
0.99846
0.99889
0.99921
0.99944
0.52790
0.56749
0.60642
0.64431
0.68082
0.71566
0.74857
0.77935
0.80785
0.83398
0.85769
0.87900
0.89796
0.91466
0.92922
0.94179
0.95254
0.96164
0.96926
0.97558
0.98077
0.98500
0.98840
0.99111
0.99324
0.99492
0.99621
0.99720
0.99795
0.99851
0.99893
0.99924
0.99946
0.53188
0.57142
0.61026
0.64803
0.68439
0.71904
0.75175
0.78230
0.81057
0.83646
0.85993
0.88100
0.89973
0.91621
0.93056
0.94295
0.95352
0.96246
0.96995
0.97615
0.98124
0.98537
0.98870
0.99134
0.99343
0.99506
0.99632
0.99728
0.99801
0.99856
0.99896
0.99926
0.99948
0.53586
0.57535
0.61409
0.65173
0.68793
0.72240
0.75490
0.78524
0.81327
0.83891
0.86214
0.88298
0.90147
0.91774
0.93189
0.94408
0.95449
0.96327
0.97062
0.97670
0.98169
0.98574
0.98899
0.99158
0.99361
0.99520
0.99643
0.99736
0.99807
0.99861
0.99900
0.99929
0.99950
564
D.2
NORMAL ERROR DISTRIBUTION CURVE AND OTHER STATISTICAL TABLES
OTHER STATISTICAL TABLES
Three often-used statistical tables whose use and interpretation are discussed
in detail in Chapters 4 and 5 are given below. The equations used to generate
each table are also presented.
D.2.1 2 Distribution
Chi square is a density function for the distribution of sample variances computed from sets with degrees of freedom selected for a population. The use
of this distribution to construct confidence intervals for the population variance and to perform hypothesis testing involving the population variance are
discussed in detail in Chapter 4. The 2 distribution is illustrated in Figure
D.2.
The 2 distribution critical values given in Table D.2 were generated using
the following function. (Critical 2 values for both tails of the distribution
were derived with a program using numerical integration routines similar to
those used in STATS.)
冕
2
␣⫽
0
1
u(⫺2) / 2e⫺u / 2 du
2 / 2⌫( /2)
where v is the degrees of freedom and ⌫ is the gamma function, which is
defined as
冕
⬁
⌫() ⫽
u⫺1e⫺u du
0
It is computed as ⌫() ⫽ ( ⫺ 1)! ⫽ ( ⫺ 1)( ⫺ 2)( ⫺ 3) 䡠 䡠 䡠 (3)(2)(1).
Figure D.2 2 distribution.
TABLE D.2 Critical Values for the 2 Distribution
␣→
0.999
0.995
0.990
0.975
0.950
0.900
0.500
0.100
0.050
0.025
0.010
0.005
0.001
0.000002
0.002
0.02
0.09
0.21
0.38
0.60
0.86
1.15
1.48
1.83
2.21
2.62
3.04
3.48
3.94
4.42
4.91
5.41
5.92
6.45
6.98
7.53
8.09
8.65
9.22
9.80
10.39
10.99
11.59
14.69
17.92
24.67
31.74
77.76
0.000039
0.01
0.07
0.21
0.41
0.68
0.99
1.34
1.74
2.16
2.60
3.07
3.57
4.08
4.60
5.14
5.70
6.27
6.84
7.43
8.03
8.64
9.26
9.89
10.52
11.16
11.81
12.46
13.12
13.79
17.19
20.71
27.99
35.53
83.85
0.000157
0.02
0.12
0.30
0.55
0.87
1.24
1.65
2.09
2.56
3.05
3.57
4.11
4.66
5.23
5.81
6.41
7.02
7.63
8.26
8.90
9.54
10.20
10.86
11.52
12.20
12.88
13.56
14.26
14.95
18.51
22.16
29.71
37.48
86.92
0.000982
0.05
0.22
0.48
0.83
1.24
1.69
2.18
2.70
3.25
3.82
4.40
5.01
5.63
6.26
6.91
7.56
8.23
8.91
9.59
10.28
10.98
11.69
12.40
13.12
13.84
14.57
15.31
16.05
16.79
20.57
24.43
32.36
40.48
91.57
0.004
0.10
0.35
0.71
1.15
1.64
2.17
2.73
3.33
3.94
4.58
5.23
5.89
6.57
7.26
7.96
8.67
9.39
10.12
10.85
11.59
12.34
13.09
13.85
14.61
15.38
16.15
16.93
17.71
18.49
22.47
26.51
34.76
43.19
95.70
0.016
0.21
0.58
1.06
1.61
2.20
2.83
3.49
4.17
4.87
5.58
6.30
7.04
7.79
8.55
9.31
10.09
10.86
11.65
12.44
13.24
14.04
14.85
15.66
16.47
17.29
18.11
18.94
19.77
20.60
24.80
29.05
37.69
46.46
100.62
0.455
1.39
2.37
3.36
4.35
5.35
6.35
7.34
8.34
9.34
10.34
11.34
12.34
13.34
14.34
15.34
16.34
17.34
18.34
19.34
20.34
21.34
22.34
23.34
24.34
25.34
26.34
27.34
28.34
29.34
34.34
39.34
49.33
59.33
119.33
2.705
4.61
6.25
7.78
9.24
10.64
12.02
13.36
14.68
15.99
17.28
18.55
19.81
21.06
22.31
23.54
24.77
25.99
27.20
28.41
29.62
30.81
32.01
33.20
34.38
35.56
36.74
37.92
39.09
40.26
46.06
51.81
63.17
74.40
140.23
3.841
5.99
7.82
9.49
11.07
12.59
14.07
15.51
16.92
18.31
19.68
21.03
22.36
23.68
25.00
26.30
27.59
28.87
30.14
31.41
32.67
33.92
35.17
36.42
37.65
38.89
40.11
41.34
42.56
43.77
49.80
55.76
67.50
79.08
146.57
5.023
7.38
9.35
11.14
12.83
14.45
16.01
17.53
19.02
20.48
21.92
23.34
24.74
26.12
27.49
28.85
30.19
31.53
32.85
34.17
35.48
36.78
38.08
39.36
40.65
41.92
43.19
44.46
45.72
46.98
53.20
59.34
71.42
83.30
152.21
6.634
9.21
11.34
13.28
15.09
16.81
18.48
20.09
21.67
23.21
24.72
26.22
27.69
29.14
30.58
32.00
33.41
34.81
36.19
37.57
38.93
40.29
41.64
42.98
44.31
45.64
46.96
48.28
49.59
50.89
57.34
63.69
76.15
88.38
158.95
7.877
10.60
12.84
14.86
16.75
18.55
20.28
21.96
23.59
25.19
26.76
28.30
29.82
31.32
32.80
34.27
35.72
37.16
38.58
40.00
41.40
42.80
44.18
45.56
46.93
48.29
49.64
50.99
52.34
53.67
60.27
66.77
79.49
91.95
163.65
10.81
13.81
16.26
18.47
20.51
22.46
24.32
26.12
27.88
29.59
31.26
32.91
34.53
36.12
37.70
39.25
40.79
42.31
43.82
45.31
46.80
48.27
49.73
51.18
52.62
54.05
55.48
56.89
58.30
59.70
66.62
73.40
86.66
99.61
173.6
↓
565
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
35
40
50
60
120
566
NORMAL ERROR DISTRIBUTION CURVE AND OTHER STATISTICAL TABLES
D.2.2 t Distribution
The t distribution (also known as Student’s distribution), shown in Figure D.3,
is used to derive confidence intervals for the population mean when the sample set is small. It is also used in hypothesis testing to check the validity of
a sample mean against a population mean. The uses for this distribution are
discussed in greater detail in Chapter 4.
The t distribution tables were generated using the following function. (Critical t values for the upper tail of the distribution were derived with a program
using numerical integration routines similar to those available in STATS.)
⌫( ⫹ 1)/2
x
冕 兹
1⫹ 冊
冉
⌫( /2)
2
␣⫽
⫺(⫹1) / 2
dx
where ⌫ is the gamma function as defined in Section D.2.1 and is the
degrees of freedom in the function. In Table D.3, critical values of t are listed
that are required to achieve the percentage points listed in the top row. The
distribution is symmetrical, and thus
F(⫺t) ⫽ 1 ⫺ F(t)
Figure D.3 t distribution.
TABLE D.3 Critical Values for the t Distribution
␣→
0.400
0.350
0.300
0.250
0.200
0.150
0.100
0.050
0.025
0.010
0.005
0.001
0.0005
0.325
0.289
0.277
0.271
0.267
0.265
0.263
0.262
0.261
0.260
0.260
0.259
0.259
0.258
0.258
0.258
0.257
0.257
0.257
0.257
0.257
0.256
0.256
0.256
0.256
0.256
0.256
0.256
0.256
0.256
0.255
0.255
0.254
0.254
0.253
0.510
0.445
0.424
0.414
0.408
0.404
0.402
0.399
0.398
0.397
0.396
0.395
0.394
0.393
0.393
0.392
0.392
0.392
0.391
0.391
0.391
0.390
0.390
0.390
0.390
0.390
0.389
0.389
0.389
0.389
0.388
0.388
0.387
0.386
0.385
0.727
0.617
0.584
0.569
0.559
0.553
0.549
0.546
0.543
0.542
0.540
0.539
0.538
0.537
0.536
0.535
0.534
0.534
0.533
0.533
0.532
0.532
0.532
0.531
0.531
0.531
0.531
0.530
0.530
0.530
0.529
0.529
0.527
0.526
0.525
1.000
0.816
0.765
0.741
0.727
0.718
0.711
0.706
0.703
0.700
0.697
0.695
0.694
0.692
0.691
0.690
0.689
0.688
0.688
0.687
0.686
0.686
0.685
0.685
0.684
0.684
0.684
0.683
0.683
0.683
0.682
0.681
0.679
0.677
0.675
1.376
1.061
0.978
0.941
0.920
0.906
0.896
0.889
0.883
0.879
0.876
0.873
0.870
0.868
0.866
0.865
0.863
0.862
0.861
0.860
0.859
0.858
0.858
0.857
0.856
0.856
0.855
0.855
0.854
0.854
0.852
0.851
0.848
0.845
0.842
1.963
1.386
1.250
1.190
1.156
1.134
1.119
1.108
1.100
1.093
1.088
1.083
1.079
1.076
1.074
1.071
1.069
1.067
1.066
1.064
1.063
1.061
1.060
1.059
1.058
1.058
1.057
1.056
1.055
1.055
1.052
1.050
1.045
1.041
1.037
3.078
1.886
1.638
1.533
1.476
1.440
1.415
1.397
1.383
1.372
1.363
1.356
1.350
1.345
1.341
1.337
1.333
1.330
1.328
1.325
1.323
1.321
1.319
1.318
1.316
1.315
1.314
1.313
1.311
1.310
1.306
1.303
1.296
1.289
1.282
6.314
2.920
2.353
2.132
2.015
1.943
1.895
1.860
1.833
1.812
1.796
1.782
1.771
1.761
1.753
1.746
1.740
1.734
1.729
1.725
1.721
1.717
1.714
1.711
1.708
1.706
1.703
1.701
1.699
1.697
1.690
1.684
1.671
1.658
1.645
12.705
4.303
3.183
2.776
2.571
2.447
2.365
2.306
2.262
2.228
2.201
2.179
2.160
2.145
2.131
2.120
2.110
2.101
2.093
2.086
2.080
2.074
2.069
2.064
2.060
2.056
2.052
2.048
2.045
2.042
2.030
2.021
2.000
1.980
1.960
31.816
6.964
4.541
3.748
3.365
3.143
2.998
2.896
2.821
2.764
2.718
2.681
2.650
2.624
2.602
2.583
2.567
2.552
2.539
2.528
2.518
2.508
2.500
2.492
2.485
2.479
2.473
2.467
2.462
2.457
2.438
2.423
2.390
2.358
2.326
63.639
9.925
5.842
4.604
4.032
3.708
3.500
3.356
3.250
3.169
3.106
3.055
3.012
2.977
2.947
2.921
2.898
2.878
2.861
2.845
2.831
2.819
2.807
2.797
2.787
2.779
2.771
2.763
2.756
2.750
2.724
2.704
2.660
2.617
2.576
318.20
22.401
10.216
6.897
5.895
5.208
4.785
4.510
4.304
4.149
4.029
3.933
3.854
3.789
3.734
3.688
3.647
3.611
3.580
3.553
3.528
3.506
3.486
3.467
3.451
3.435
3.421
3.409
3.397
3.385
3.340
3.307
3.232
3.160
3.291
636.18
31.579
12.954
8.610
6.880
5.961
5.408
5.041
4.781
4.605
4.452
4.329
4.230
4.148
4.079
4.021
3.970
3.926
3.887
3.853
3.822
3.795
3.770
3.748
3.727
3.708
3.691
3.675
3.661
3.647
3.592
3.552
3.461
3.374
3.300
↓
567
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
35
40
60
120
⬁
568
NORMAL ERROR DISTRIBUTION CURVE AND OTHER STATISTICAL TABLES
D.2.3 F Distribution
This F distribution (also known as the Fisher distribution), shown in Figure
D.4, is used to derive confidence intervals for the ratio of two population
variances. It is also used in hypothesis testing for this ratio. Uses for this
distribution are discussed in Chapter 4.
Critical F values for the upper tail of the distribution were derived with a
program using numerical integration routines similar to those used in STATS.
The tables were generated using the following function:
冕
⬁
␣⫽
F
⌫(1 ⫹ 2)/2
⌫(1 /2)⌫(2 /2)
冉冊
1
2
1 / 2
x(1⫺2) / 2
dx
1 ⫹ (1 / 2)x(1⫹2) / 2
where ⌫ is the gamma function as defined in Section D.2.1, 1 the numerator
degrees of freedom, and 2 the denominator degrees of freedom.
For critical values in the lower tail of the distribution, the following relationship can be used in conjunction with the tabular values given in Table
D.4.
F␣,1,2 ⫽
1
F1⫺␣,2,1
Figure D.4 F distribution.
TABLE D.4 Critical Values for the F Distribution
␣ ⴝ 0.20
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
24
30
40
60
120
14.77
4.38
2.98
2.46
2.19
2.03
1.92
1.84
1.78
1.73
1.69
1.66
1.64
1.62
1.60
1.58
1.57
1.55
1.54
1.53
1.52
1.51
1.51
1.50
1.49
1.49
1.48
1.48
1.47
1.47
1.42
1.41
1.39
1.37
14.90
4.40
2.98
2.46
2.18
2.02
1.91
1.83
1.76
1.72
1.68
1.65
1.62
1.60
1.58
1.56
1.55
1.53
1.52
1.51
1.50
1.49
1.49
1.48
1.47
1.47
1.46
1.46
1.45
1.45
1.39
1.38
1.37
1.35
15.04
4.42
2.98
2.45
2.18
2.01
1.89
1.81
1.75
1.70
1.66
1.63
1.60
1.58
1.56
1.54
1.53
1.51
1.50
1.49
1.48
1.47
1.46
1.46
1.45
1.44
1.44
1.43
1.43
1.42
1.37
1.35
1.34
1.32
15.17
4.43
2.98
2.44
2.17
2.00
1.88
1.80
1.73
1.68
1.64
1.61
1.58
1.56
1.54
1.52
1.50
1.49
1.48
1.47
1.46
1.45
1.44
1.43
1.42
1.42
1.41
1.41
1.40
1.39
1.34
1.32
1.31
1.29
15.24
4.44
2.98
2.44
2.16
1.99
1.87
1.79
1.72
1.67
1.63
1.60
1.57
1.55
1.53
1.51
1.49
1.48
1.46
1.45
1.44
1.43
1.42
1.42
1.41
1.40
1.40
1.39
1.39
1.38
1.32
1.31
1.29
1.27
15.31
4.45
2.98
2.44
2.16
1.98
1.86
1.78
1.71
1.66
1.62
1.59
1.56
1.53
1.51
1.49
1.48
1.46
1.45
1.44
1.43
1.42
1.41
1.40
1.39
1.39
1.38
1.37
1.37
1.36
1.30
1.29
1.27
1.25
15.37
4.46
2.98
2.44
2.15
1.98
1.86
1.77
1.70
1.65
1.61
1.58
1.55
1.52
1.50
1.48
1.46
1.45
1.44
1.42
1.41
1.40
1.39
1.39
1.38
1.37
1.36
1.36
1.35
1.35
1.28
1.27
1.25
1.23
15.44
15.51
4.46
4.47
2.98
2.98
2.43
2.43
2.15
2.14
1.97
1.96
1.85
1.84
1.76
1.75
1.69
1.68
1.64
1.63
1.60
1.59
1.56
1.55
1.53
1.52
1.51
1.49
1.49
1.47
1.47
1.45
1.45
1.43
1.43
1.42
1.42
1.40
1.41
1.39
1.40
1.38
1.39
1.37
1.38
1.36
1.37
1.35
1.36
1.34
1.35
1.33
1.35
1.33
1.34
1.32
1.33
1.31
1.33
1.31
1.26
1.24
1.24
1.22
1.22
1.19
1.20
1.17
(continues)
2↓
569
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
50
60
80
120
9.47
3.56
2.68
2.35
2.18
2.07
2.00
1.95
1.91
1.88
1.86
1.84
1.82
1.81
1.80
1.79
1.78
1.77
1.76
1.76
1.75
1.75
1.74
1.74
1.73
1.73
1.73
1.72
1.72
1.72
1.69
1.68
1.67
1.66
12.00
4.00
2.89
2.47
2.26
2.13
2.04
1.98
1.93
1.90
1.87
1.85
1.83
1.81
1.80
1.78
1.77
1.76
1.75
1.75
1.74
1.73
1.73
1.72
1.72
1.71
1.71
1.71
1.70
1.70
1.66
1.65
1.64
1.63
13.06
4.16
2.94
2.48
2.25
2.11
2.02
1.95
1.90
1.86
1.83
1.80
1.78
1.76
1.75
1.74
1.72
1.71
1.70
1.70
1.69
1.68
1.68
1.67
1.66
1.66
1.66
1.65
1.65
1.64
1.60
1.59
1.58
1.57
13.64
4.24
2.96
2.48
2.24
2.09
1.99
1.92
1.87
1.83
1.80
1.77
1.75
1.73
1.71
1.70
1.68
1.67
1.66
1.65
1.65
1.64
1.63
1.63
1.62
1.62
1.61
1.61
1.60
1.60
1.56
1.55
1.53
1.52
14.01
4.28
2.97
2.48
2.23
2.08
1.97
1.90
1.85
1.80
1.77
1.74
1.72
1.70
1.68
1.67
1.65
1.64
1.63
1.62
1.61
1.61
1.60
1.59
1.59
1.58
1.58
1.57
1.57
1.57
1.52
1.51
1.50
1.48
14.26
4.32
2.97
2.47
2.22
2.06
1.96
1.88
1.83
1.78
1.75
1.72
1.69
1.67
1.66
1.64
1.63
1.62
1.61
1.60
1.59
1.58
1.57
1.57
1.56
1.56
1.55
1.55
1.54
1.54
1.49
1.48
1.47
1.45
14.44
4.34
2.97
2.47
2.21
2.05
1.94
1.87
1.81
1.77
1.73
1.70
1.68
1.65
1.64
1.62
1.61
1.60
1.58
1.58
1.57
1.56
1.55
1.55
1.54
1.53
1.53
1.52
1.52
1.52
1.47
1.46
1.44
1.43
14.58
4.36
2.98
2.47
2.20
2.04
1.93
1.86
1.80
1.75
1.72
1.69
1.66
1.64
1.62
1.61
1.59
1.58
1.57
1.56
1.55
1.54
1.53
1.53
1.52
1.52
1.51
1.51
1.50
1.50
1.45
1.44
1.42
1.41
14.68
4.37
2.98
2.46
2.20
2.03
1.93
1.85
1.79
1.74
1.70
1.67
1.65
1.63
1.61
1.59
1.58
1.56
1.55
1.54
1.53
1.53
1.52
1.51
1.51
1.50
1.49
1.49
1.49
1.48
1.43
1.42
1.41
1.39
570
TABLE D.4 (Continued )
␣ ⴝ 0.10
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
24
30
40
60
120
60.19
9.39
5.23
3.92
3.30
2.94
2.70
2.54
2.42
2.32
2.25
2.19
2.14
2.10
2.06
2.03
2.00
1.98
1.96
1.94
1.92
1.90
1.89
1.88
1.87
1.86
1.85
1.84
1.83
1.82
1.73
1.71
1.68
1.65
60.70
9.41
5.22
3.90
3.27
2.90
2.67
2.50
2.38
2.28
2.21
2.15
2.10
2.05
2.02
1.99
1.96
1.93
1.91
1.89
1.87
1.86
1.84
1.83
1.82
1.81
1.80
1.79
1.78
1.77
1.68
1.66
1.63
1.60
61.21
9.42
5.20
3.87
3.24
2.87
2.63
2.46
2.34
2.24
2.17
2.10
2.05
2.01
1.97
1.94
1.91
1.89
1.86
1.84
1.83
1.81
1.80
1.78
1.77
1.76
1.75
1.74
1.73
1.72
1.63
1.60
1.57
1.55
61.73
9.44
5.18
3.84
3.21
2.84
2.59
2.42
2.30
2.20
2.12
2.06
2.01
1.96
1.92
1.89
1.86
1.84
1.81
1.79
1.78
1.76
1.74
1.73
1.72
1.71
1.70
1.69
1.68
1.67
1.57
1.54
1.51
1.48
61.99
9.45
5.18
3.83
3.19
2.82
2.58
2.40
2.28
2.18
2.10
2.04
1.98
1.94
1.90
1.87
1.84
1.81
1.79
1.77
1.75
1.73
1.72
1.70
1.69
1.68
1.67
1.66
1.65
1.64
1.54
1.51
1.48
1.45
62.26
9.46
5.17
3.82
3.17
2.80
2.56
2.38
2.25
2.16
2.08
2.01
1.96
1.91
1.87
1.84
1.81
1.78
1.76
1.74
1.72
1.70
1.69
1.67
1.66
1.65
1.64
1.63
1.62
1.61
1.50
1.48
1.44
1.41
62.52
9.47
5.16
3.80
3.16
2.78
2.54
2.36
2.23
2.13
2.05
1.99
1.93
1.89
1.85
1.81
1.78
1.75
1.73
1.71
1.69
1.67
1.66
1.64
1.63
1.61
1.60
1.59
1.58
1.57
1.46
1.44
1.40
1.37
62.79
9.47
5.15
3.79
3.14
2.76
2.51
2.34
2.21
2.11
2.03
1.96
1.90
1.86
1.82
1.78
1.75
1.72
1.70
1.68
1.66
1.64
1.62
1.61
1.59
1.58
1.57
1.56
1.55
1.54
1.42
1.40
1.36
1.32
63.05
9.48
5.14
3.78
3.12
2.74
2.49
2.32
2.18
2.08
2.00
1.93
1.88
1.83
1.79
1.75
1.72
1.69
1.67
1.64
1.62
1.60
1.59
1.57
1.56
1.54
1.53
1.52
1.51
1.50
1.38
1.35
1.31
1.26
2↓
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
50
60
80
120
39.85
8.53
5.54
4.54
4.06
3.78
3.59
3.46
3.36
3.28
3.23
3.18
3.14
3.10
3.07
3.05
3.03
3.01
2.99
2.97
2.96
2.95
2.94
2.93
2.92
2.91
2.90
2.89
2.89
2.88
2.81
2.79
2.77
2.75
49.49
9.00
5.46
4.32
3.78
3.46
3.26
3.11
3.01
2.92
2.86
2.81
2.76
2.73
2.70
2.67
2.64
2.62
2.61
2.59
2.57
2.56
2.55
2.54
2.53
2.52
2.51
2.50
2.50
2.49
2.41
2.39
2.37
2.35
53.59
9.16
5.39
4.19
3.62
3.29
3.07
2.92
2.81
2.73
2.66
2.61
2.56
2.52
2.49
2.46
2.44
2.42
2.40
2.38
2.36
2.35
2.34
2.33
2.32
2.31
2.30
2.29
2.28
2.28
2.20
2.18
2.15
2.13
55.83
9.24
5.34
4.11
3.52
3.18
2.96
2.81
2.69
2.61
2.54
2.48
2.43
2.39
2.36
2.33
2.31
2.29
2.27
2.25
2.23
2.22
2.21
2.19
2.18
2.17
2.17
2.16
2.15
2.14
2.06
2.04
2.02
1.99
57.23
9.29
5.31
4.05
3.45
3.11
2.88
2.73
2.61
2.52
2.45
2.39
2.35
2.31
2.27
2.24
2.22
2.20
2.18
2.16
2.14
2.13
2.11
2.10
2.09
2.08
2.07
2.06
2.06
2.05
1.97
1.95
1.92
1.90
58.20
9.33
5.28
4.01
3.40
3.05
2.83
2.67
2.55
2.46
2.39
2.33
2.28
2.24
2.21
2.18
2.15
2.13
2.11
2.09
2.08
2.06
2.05
2.04
2.02
2.01
2.00
2.00
1.99
1.98
1.90
1.87
1.85
1.82
58.90
9.35
5.27
3.98
3.37
3.01
2.78
2.62
2.51
2.41
2.34
2.28
2.23
2.19
2.16
2.13
2.10
2.08
2.06
2.04
2.02
2.01
1.99
1.98
1.97
1.96
1.95
1.94
1.93
1.93
1.84
1.82
1.79
1.77
59.43
9.37
5.25
3.95
3.34
2.98
2.75
2.59
2.47
2.38
2.30
2.24
2.20
2.15
2.12
2.09
2.06
2.04
2.02
2.00
1.98
1.97
1.95
1.94
1.93
1.92
1.91
1.90
1.89
1.88
1.80
1.77
1.75
1.72
59.85
9.38
5.24
3.94
3.32
2.96
2.72
2.56
2.44
2.35
2.27
2.21
2.16
2.12
2.09
2.06
2.03
2.00
1.98
1.96
1.95
1.93
1.92
1.91
1.89
1.88
1.87
1.87
1.86
1.85
1.76
1.74
1.71
1.68
␣ ⴝ 0.05
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
24
30
40
161.4
18.51
10.13
7.71
6.61
5.99
5.59
5.32
5.12
4.96
4.84
4.75
4.67
4.60
4.54
4.49
4.45
4.41
4.38
4.35
4.32
4.30
4.28
4.26
4.24
4.22
4.21
4.20
4.18
4.17
4.03
4.00
3.96
3.92
199.5
19.00
9.55
6.94
5.79
5.14
4.74
4.46
4.26
4.10
3.98
3.89
3.81
3.74
3.68
3.63
3.59
3.55
3.52
3.49
3.47
3.44
3.42
3.40
3.39
3.37
3.35
3.34
3.33
3.32
3.18
3.15
3.11
3.07
215.7
19.16
9.28
6.59
5.41
4.76
4.35
4.07
3.86
3.71
3.59
3.49
3.41
3.34
3.29
3.24
3.20
3.16
3.13
3.10
3.07
3.05
3.03
3.01
2.99
2.98
2.96
2.95
2.93
2.92
2.79
2.76
2.72
2.68
224.6
19.25
9.12
6.39
5.19
4.53
4.12
3.84
3.63
3.48
3.36
3.26
3.18
3.11
3.06
3.01
2.96
2.93
2.90
2.87
2.84
2.82
2.80
2.78
2.76
2.74
2.73
2.71
2.70
2.69
2.56
2.53
2.49
2.45
230.2
19.30
9.01
6.26
5.05
4.39
3.97
3.69
3.48
3.33
3.20
3.11
3.03
2.96
2.90
2.85
2.81
2.77
2.74
2.71
2.68
2.66
2.64
2.62
2.60
2.59
2.57
2.56
2.55
2.53
2.40
2.37
2.33
2.29
234.0
19.33
8.94
6.16
4.95
4.28
3.87
3.58
3.37
3.22
3.09
3.00
2.92
2.85
2.79
2.74
2.70
2.66
2.63
2.60
2.57
2.55
2.53
2.51
2.49
2.47
2.46
2.45
2.43
2.42
2.29
2.25
2.21
2.18
236.8
19.35
8.89
6.09
4.88
4.21
3.79
3.50
3.29
3.14
3.01
2.91
2.83
2.76
2.71
2.66
2.61
2.58
2.54
2.51
2.49
2.46
2.44
2.42
2.40
2.39
2.37
2.36
2.35
2.33
2.20
2.17
2.13
2.09
238.9
19.37
8.85
6.04
4.82
4.15
3.73
3.44
3.23
3.07
2.95
2.85
2.77
2.70
2.64
2.59
2.55
2.51
2.48
2.45
2.42
2.40
2.37
2.36
2.34
2.32
2.31
2.29
2.28
2.27
2.13
2.10
2.06
2.02
240.5
19.38
8.81
6.00
4.77
4.10
3.68
3.39
3.18
3.02
2.90
2.80
2.71
2.65
2.59
2.54
2.49
2.46
2.42
2.39
2.37
2.34
2.32
2.30
2.28
2.27
2.25
2.24
2.22
2.21
2.07
2.04
2.00
1.96
241.9
19.40
8.79
5.96
4.74
4.06
3.64
3.35
3.14
2.98
2.85
2.75
2.67
2.60
2.54
2.49
2.45
2.41
2.38
2.35
2.32
2.30
2.27
2.25
2.24
2.22
2.20
2.19
2.18
2.16
2.03
1.99
1.95
1.91
243.9
19.41
8.74
5.91
4.68
4.00
3.57
3.28
3.07
2.91
2.79
2.69
2.60
2.53
2.48
2.42
2.38
2.34
2.31
2.28
2.25
2.23
2.20
2.18
2.16
2.15
2.13
2.12
2.10
2.09
1.95
1.92
1.88
1.83
245.9
19.43
8.70
5.86
4.62
3.94
3.51
3.22
3.01
2.85
2.72
2.62
2.53
2.46
2.40
2.35
2.31
2.27
2.23
2.20
2.18
2.15
2.13
2.11
2.09
2.07
2.06
2.04
2.03
2.01
1.87
1.84
1.79
1.75
248
19.45
8.66
5.80
4.56
3.87
3.44
3.15
2.94
2.77
2.65
2.54
2.46
2.39
2.33
2.28
2.23
2.19
2.16
2.12
2.10
2.07
2.05
2.03
2.01
1.99
1.97
1.96
1.94
1.93
1.78
1.75
1.70
1.66
249
19.45
8.64
5.77
4.53
3.84
3.41
3.12
2.90
2.74
2.61
2.51
2.42
2.35
2.29
2.24
2.19
2.15
2.11
2.08
2.05
2.03
2.01
1.98
1.96
1.95
1.93
1.91
1.90
1.89
1.74
1.70
1.65
1.61
250
19.46
8.62
5.75
4.50
3.81
3.38
3.08
2.86
2.70
2.57
2.47
2.38
2.31
2.25
2.19
2.15
2.11
2.07
2.04
2.01
1.98
1.96
1.94
1.92
1.90
1.88
1.87
1.85
1.84
1.69
1.65
1.60
1.55
251
19.47
8.59
5.72
4.46
3.77
3.34
3.04
2.83
2.66
2.53
2.43
2.34
2.27
2.20
2.15
2.10
2.06
2.03
1.99
1.96
1.94
1.91
1.89
1.87
1.85
1.84
1.82
1.81
1.79
1.63
1.59
1.54
1.50
60
120
2↓
571
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
50
60
80
120
252
253.2
19.48
19.49
8.57
8.55
5.69
5.66
4.43
4.40
3.74
3.70
3.30
3.27
3.01
2.97
2.79
2.75
2.62
2.58
2.49
2.45
2.38
2.34
2.30
2.25
2.22
2.18
2.16
2.11
2.11
2.06
2.06
2.01
2.02
1.97
1.98
1.93
1.95
1.90
1.92
1.87
1.89
1.84
1.86
1.81
1.84
1.79
1.82
1.77
1.80
1.75
1.79
1.73
1.77
1.71
1.75
1.70
1.74
1.68
1.58
1.51
1.53
1.47
1.48
1.41
1.43
1.35
(continues)
572
TABLE D.4 (Continued )
␣ ⴝ 0.025
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
647.8
38.51
17.44
12.22
10.01
8.81
8.07
7.57
7.21
6.94
6.72
6.55
6.41
6.30
6.20
6.11
6.04
5.98
5.92
5.87
5.83
5.79
5.75
5.72
5.69
5.66
5.63
5.61
5.59
5.57
5.34
5.29
5.22
5.15
799.5
39.00
16.04
10.65
8.43
7.26
6.54
6.06
5.71
5.46
5.26
5.10
4.97
4.86
4.76
4.69
4.62
4.56
4.51
4.46
4.42
4.38
4.35
4.32
4.29
4.27
4.24
4.22
4.20
4.18
3.97
3.93
3.86
3.80
864.2
39.17
15.44
9.98
7.76
6.60
5.89
5.42
5.08
4.83
4.63
4.47
4.35
4.24
4.15
4.08
4.01
3.95
3.90
3.86
3.82
3.78
3.75
3.72
3.69
3.67
3.65
3.63
3.61
3.59
3.39
3.34
3.28
3.23
899.6
39.25
15.10
9.60
7.39
6.23
5.52
5.05
4.72
4.47
4.28
4.12
4.00
3.89
3.80
3.73
3.66
3.61
3.56
3.51
3.48
3.44
3.41
3.38
3.35
3.33
3.31
3.29
3.27
3.25
3.05
3.01
2.95
2.89
921.8
39.30
14.88
9.36
7.15
5.99
5.29
4.82
4.48
4.24
4.04
3.89
3.77
3.66
3.58
3.50
3.44
3.38
3.33
3.29
3.25
3.22
3.18
3.15
3.13
3.10
3.08
3.06
3.04
3.03
2.83
2.79
2.73
2.67
937.1
39.33
14.73
9.20
6.98
5.82
5.12
4.65
4.32
4.07
3.88
3.73
3.60
3.50
3.41
3.34
3.28
3.22
3.17
3.13
3.09
3.05
3.02
2.99
2.97
2.94
2.92
2.90
2.88
2.87
2.67
2.63
2.57
2.52
948.2
39.36
14.62
9.07
6.85
5.70
4.99
4.53
4.20
3.95
3.76
3.61
3.48
3.38
3.29
3.22
3.16
3.10
3.05
3.01
2.97
2.93
2.90
2.87
2.85
2.82
2.80
2.78
2.76
2.75
2.55
2.51
2.45
2.39
956.7
39.37
14.54
8.98
6.76
5.60
4.90
4.43
4.10
3.85
3.66
3.51
3.39
3.29
3.20
3.12
3.06
3.01
2.96
2.91
2.87
2.84
2.81
2.78
2.75
2.73
2.71
2.69
2.67
2.65
2.46
2.41
2.35
2.30
963.3
39.39
14.47
8.90
6.68
5.52
4.82
4.36
4.03
3.78
3.59
3.44
3.31
3.21
3.12
3.05
2.98
2.93
2.88
2.84
2.80
2.76
2.73
2.70
2.68
2.65
2.63
2.61
2.59
2.57
2.38
2.33
2.28
2.22
968.6
39.40
14.42
8.84
6.62
5.46
4.76
4.30
3.96
3.72
3.53
3.37
3.25
3.15
3.06
2.99
2.92
2.87
2.82
2.77
2.73
2.70
2.67
2.64
2.61
2.59
2.57
2.55
2.53
2.51
2.32
2.27
2.21
2.16
976.7
39.41
14.34
8.75
6.52
5.37
4.67
4.20
3.87
3.62
3.43
3.28
3.15
3.05
2.96
2.89
2.82
2.77
2.72
2.68
2.64
2.60
2.57
2.54
2.51
2.49
2.47
2.45
2.43
2.41
2.22
2.17
2.11
2.05
984.9
39.43
14.25
8.66
6.43
5.27
4.57
4.10
3.77
3.52
3.33
3.18
3.05
2.95
2.86
2.79
2.72
2.67
2.62
2.57
2.53
2.50
2.47
2.44
2.41
2.39
2.36
2.34
2.32
2.31
2.11
2.06
2.00
1.94
993.1
39.45
14.17
8.56
6.33
5.17
4.47
4.00
3.67
3.42
3.23
3.07
2.95
2.84
2.76
2.68
2.62
2.56
2.51
2.46
2.42
2.39
2.36
2.33
2.30
2.28
2.25
2.23
2.21
2.20
1.99
1.94
1.88
1.82
24
30
40
60
120
2↓
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
50
60
80
120
997.2 1001
1006
1010
1014
39.46
39.46
39.47
39.48
39.48
14.12
14.08
14.04
13.99
13.95
8.51
8.46
8.41
8.36
8.31
6.28
6.23
6.18
6.12
6.07
5.12
5.07
5.01
4.96
4.90
4.41
4.36
4.31
4.25
4.20
3.95
3.89
3.84
3.78
3.73
3.61
3.56
3.51
3.45
3.39
3.37
3.31
3.26
3.20
3.14
3.17
3.12
3.06
3.00
2.94
3.02
2.96
2.91
2.85
2.79
2.89
2.84
2.78
2.72
2.66
2.79
2.73
2.67
2.61
2.55
2.70
2.64
2.59
2.52
2.46
2.63
2.57
2.51
2.45
2.38
2.56
2.50
2.44
2.38
2.32
2.50
2.44
2.38
2.32
2.26
2.45
2.39
2.33
2.27
2.20
2.41
2.35
2.29
2.22
2.16
2.37
2.31
2.25
2.18
2.11
2.33
2.27
2.21
2.14
2.08
2.30
2.24
2.18
2.11
2.04
2.27
2.21
2.15
2.08
2.01
2.24
2.18
2.12
2.05
1.98
2.22
2.16
2.09
2.03
1.95
2.19
2.13
2.07
2.00
1.93
2.17
2.11
2.05
1.98
1.91
2.15
2.09
2.03
1.96
1.89
2.14
2.07
2.01
1.94
1.87
1.93
1.87
1.80
1.72
1.64
1.88
1.82
1.74
1.67
1.58
1.82
1.75
1.68
1.60
1.51
1.76
1.69
1.61
1.53
1.43
␣ ⴝ 0.01
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
24
30
40
60
120
2↓
573
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
50
60
80
120
4052
5000
5403
5625
5764
5859
5928
5982
6022
6056
6106
6157
6209
6235
6261
6287
6313
6339
98.5
99.0
99.2
99.2
99.3
99.3
99.4
99.4
99.4
99.4
99.4
99.4
99.4
99.5
99.5
99.5
99.5
99.5
34.1
30.8
29.5
28.7
28.2
27.9
27.7
27.5
27.3
27.2
27.1
26.9
26.7
26.6
26.5
26.4
26.3
26.2
21.2
18.0
16.7
16.0
15.5
15.2
15.0
14.8
14.7
14.6
14.4
14.2
14.0
13.9
13.8
13.7
13.7
13.6
16.25
13.27
12.06
11.39
10.97
10.67
10.46
10.29
10.16
10.05
9.89
9.72
9.55
9.47
9.38
9.29
9.20
9.11
13.74
10.92
9.78
9.15
8.75
8.47
8.26
8.10
7.98
7.87
7.72
7.56
7.40
7.31
7.23
7.14
7.06
6.97
12.24
9.55
8.45
7.85
7.46
7.19
6.99
6.84
6.72
6.62
6.47
6.31
6.16
6.07
5.99
5.91
5.82
5.74
11.26
8.65
7.59
7.01
6.63
6.37
6.18
6.03
5.91
5.81
5.67
5.52
5.36
5.28
5.20
5.12
5.03
4.95
10.56
8.02
6.99
6.42
6.06
5.80
5.61
5.47
5.35
5.26
5.11
4.96
4.81
4.73
4.65
4.57
4.48
4.40
10.04
7.56
6.55
5.99
5.64
5.39
5.20
5.06
4.94
4.85
4.71
4.56
4.41
4.33
4.25
4.17
4.08
4.00
9.64
7.21
6.22
5.67
5.32
5.07
4.89
4.74
4.63
4.54
4.40
4.25
4.10
4.02
3.94
3.86
3.78
3.69
9.33
6.93
5.95
5.41
5.06
4.82
4.64
4.50
4.39
4.30
4.16
4.01
3.86
3.78
3.70
3.62
3.54
3.45
9.07
6.70
5.74
5.21
4.86
4.62
4.44
4.30
4.19
4.10
3.96
3.82
3.66
3.59
3.51
3.43
3.34
3.25
8.86
6.51
5.56
5.04
4.69
4.46
4.28
4.14
4.03
3.94
3.80
3.66
3.51
3.43
3.35
3.27
3.18
3.09
8.68
6.36
5.42
4.89
4.56
4.32
4.14
4.00
3.89
3.80
3.67
3.52
3.37
3.29
3.21
3.13
3.05
2.96
8.53
6.23
5.29
4.77
4.44
4.20
4.03
3.89
3.78
3.69
3.55
3.41
3.26
3.18
3.10
3.02
2.93
2.84
8.40
6.11
5.18
4.67
4.34
4.10
3.93
3.79
3.68
3.59
3.46
3.31
3.16
3.08
3.00
2.92
2.83
2.75
8.28
6.01
5.09
4.58
4.25
4.01
3.84
3.71
3.60
3.51
3.37
3.23
3.08
3.00
2.92
2.84
2.75
2.66
8.18
5.93
5.01
4.50
4.17
3.94
3.77
3.63
3.52
3.43
3.30
3.15
3.00
2.92
2.84
2.76
2.67
2.58
8.09
5.85
4.94
4.43
4.10
3.87
3.70
3.56
3.46
3.37
3.23
3.09
2.94
2.86
2.78
2.69
2.61
2.52
8.01
5.78
4.87
4.37
4.04
3.81
3.64
3.51
3.40
3.31
3.17
3.03
2.88
2.80
2.72
2.64
2.55
2.46
7.94
5.72
4.82
4.31
3.99
3.76
3.59
3.45
3.35
3.26
3.12
2.98
2.83
2.75
2.67
2.58
2.50
2.40
7.88
5.66
4.76
4.26
3.94
3.71
3.54
3.41
3.30
3.21
3.07
2.93
2.78
2.70
2.62
2.54
2.45
2.35
7.82
5.61
4.72
4.22
3.90
3.67
3.50
3.36
3.26
3.17
3.03
2.89
2.74
2.66
2.58
2.49
2.40
2.31
7.77
5.57
4.68
4.18
3.85
3.63
3.46
3.32
3.22
3.13
2.99
2.85
2.70
2.62
2.54
2.45
2.36
2.27
7.72
5.53
4.64
4.14
3.82
3.59
3.42
3.29
3.18
3.09
2.96
2.81
2.66
2.58
2.50
2.42
2.33
2.23
7.67
5.49
4.60
4.11
3.78
3.56
3.39
3.26
3.15
3.06
2.93
2.78
2.63
2.55
2.47
2.38
2.29
2.20
7.63
5.45
4.57
4.07
3.75
3.53
3.36
3.23
3.12
3.03
2.90
2.75
2.60
2.52
2.44
2.35
2.26
2.17
7.60
5.42
4.54
4.04
3.73
3.50
3.33
3.20
3.09
3.00
2.87
2.73
2.57
2.49
2.41
2.33
2.23
2.14
7.56
5.39
4.51
4.02
3.70
3.47
3.30
3.17
3.07
2.98
2.84
2.70
2.55
2.47
2.39
2.30
2.21
2.11
7.17
5.06
4.20
3.72
3.41
3.19
3.02
2.89
2.78
2.70
2.56
2.42
2.27
2.18
2.10
2.01
1.91
1.80
7.08
4.98
4.13
3.65
3.34
3.12
2.95
2.82
2.72
2.63
2.50
2.35
2.20
2.12
2.03
1.94
1.84
1.73
6.96
4.88
4.04
3.56
3.26
3.04
2.87
2.74
2.64
2.55
2.42
2.27
2.12
2.03
1.94
1.85
1.75
1.63
6.85
4.79
3.95
3.48
3.17
2.96
2.79
2.66
2.56
2.47
2.34
2.19
2.03
1.95
1.86
1.76
1.66
1.53
(continues)
574
TABLE D.4 (Continued )
␣ ⴝ 0.005
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
24
30
40
60
120
16211
198.5
55.55
31.33
22.77
18.62
16.23
14.68
13.61
12.82
12.22
11.75
11.37
11.06
10.79
10.57
10.38
10.21
10.07
9.94
9.83
9.72
9.63
9.55
9.47
9.40
9.34
9.28
9.23
9.18
8.62
8.49
8.33
8.18
20000
199.0
49.80
26.28
18.31
14.54
12.40
11.04
10.10
9.43
8.91
8.51
8.19
7.92
7.70
7.51
7.35
7.21
7.09
6.99
6.89
6.81
6.73
6.66
6.60
6.54
6.49
6.44
6.39
6.35
5.90
5.79
5.66
5.54
21615
199.2
47.47
24.26
16.53
12.91
10.88
9.60
8.72
8.08
7.60
7.23
6.93
6.68
6.48
6.30
6.16
6.03
5.92
5.82
5.73
5.65
5.58
5.52
5.46
5.41
5.36
5.32
5.28
5.24
4.83
4.73
4.61
4.50
22500
199.2
46.19
23.15
15.55
12.03
10.05
8.80
7.96
7.34
6.88
6.52
6.23
6.00
5.80
5.64
5.50
5.37
5.27
5.17
5.09
5.02
4.95
4.89
4.83
4.79
4.74
4.70
4.66
4.62
4.23
4.14
4.03
3.92
23056
199.3
45.39
22.46
14.94
11.46
9.52
8.30
7.47
6.87
6.42
6.07
5.79
5.56
5.37
5.21
5.07
4.96
4.85
4.76
4.68
4.61
4.54
4.49
4.43
4.38
4.34
4.30
4.26
4.23
3.85
3.76
3.65
3.55
23437
199.4
44.84
21.97
14.51
11.07
9.16
7.95
7.13
6.54
6.10
5.76
5.48
5.26
5.07
4.91
4.78
4.66
4.56
4.47
4.39
4.32
4.26
4.20
4.15
4.10
4.06
4.02
3.98
3.95
3.58
3.49
3.39
3.28
23715
199.4
44.43
21.62
14.20
10.79
8.89
7.69
6.88
6.30
5.86
5.52
5.25
5.03
4.85
4.69
4.56
4.44
4.34
4.26
4.18
4.11
4.05
3.99
3.94
3.89
3.85
3.81
3.77
3.74
3.38
3.29
3.19
3.09
23925
199.4
44.13
21.35
13.96
10.57
8.68
7.50
6.69
6.12
5.68
5.35
5.08
4.86
4.67
4.52
4.39
4.28
4.18
4.09
4.01
3.94
3.88
3.83
3.78
3.73
3.69
3.65
3.61
3.58
3.22
3.13
3.03
2.93
24091
199.4
43.88
21.14
13.77
10.39
8.51
7.34
6.54
5.97
5.54
5.20
4.94
4.72
4.54
4.38
4.25
4.14
4.04
3.96
3.88
3.81
3.75
3.69
3.64
3.60
3.56
3.52
3.48
3.45
3.09
3.01
2.91
2.81
24224
199.4
43.69
20.97
13.62
10.25
8.38
7.21
6.42
5.85
5.42
5.09
4.82
4.60
4.42
4.27
4.14
4.03
3.93
3.85
3.77
3.70
3.64
3.59
3.54
3.49
3.45
3.41
3.38
3.34
2.99
2.90
2.80
2.71
24426
199.4
43.39
20.70
13.38
10.03
8.18
7.01
6.23
5.66
5.24
4.91
4.64
4.43
4.25
4.10
3.97
3.86
3.76
3.68
3.60
3.54
3.47
3.42
3.37
3.33
3.28
3.25
3.21
3.18
2.82
2.74
2.64
2.54
24630
199.4
43.06
20.44
13.15
9.81
7.97
6.81
6.03
5.47
5.05
4.72
4.46
4.25
4.07
3.92
3.79
3.68
3.59
3.50
3.43
3.36
3.30
3.25
3.20
3.15
3.11
3.07
3.04
3.01
2.65
2.57
2.47
2.37
24836
199.4
42.78
20.17
12.90
9.59
7.75
6.61
5.83
5.27
4.86
4.53
4.27
4.06
3.88
3.73
3.61
3.50
3.40
3.32
3.24
3.18
3.12
3.06
3.01
2.97
2.93
2.89
2.86
2.82
2.47
2.39
2.29
2.19
24940
199.5
42.62
20.03
12.78
9.47
7.64
6.50
5.73
5.17
4.76
4.43
4.17
3.96
3.79
3.64
3.51
3.40
3.31
3.22
3.15
3.08
3.02
2.97
2.92
2.87
2.83
2.79
2.76
2.73
2.37
2.29
2.19
2.09
25044
199.5
42.47
19.89
12.66
9.36
7.53
6.40
5.62
5.07
4.65
4.33
4.07
3.86
3.69
3.54
3.41
3.30
3.21
3.12
3.05
2.98
2.92
2.87
2.82
2.77
2.73
2.69
2.66
2.63
2.27
2.19
2.08
1.98
25148
199.5
42.31
19.75
12.53
9.24
7.42
6.29
5.52
4.97
4.55
4.23
3.97
3.76
3.58
3.44
3.31
3.20
3.11
3.02
2.95
2.88
2.82
2.77
2.72
2.67
2.63
2.59
2.56
2.52
2.16
2.08
1.97
1.87
25253
199.5
42.15
19.61
12.40
9.12
7.31
6.18
5.41
4.86
4.44
4.12
3.87
3.66
3.48
3.33
3.21
3.10
3.00
2.92
2.84
2.77
2.71
2.66
2.61
2.56
2.52
2.48
2.45
2.42
2.05
1.96
1.85
1.75
253591
199.5
41.99
19.47
12.27
9.00
7.19
6.06
5.30
4.75
4.34
4.01
3.76
3.55
3.37
3.22
3.10
2.99
2.89
2.81
2.73
2.66
2.60
2.55
2.50
2.45
2.41
2.37
2.33
2.30
1.93
1.83
1.72
1.61
2↓
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
50
60
80
120
␣ ⴝ 0.001
1→
1
2
3
4
5
6
7
8
9
10
12
15
20
24
30
40
60
120
405269
998.5
167.0
74.14
47.18
35.51
29.25
25.41
22.86
21.04
19.69
18.64
17.82
17.14
16.59
16.12
15.72
15.38
15.08
14.82
14.59
14.38
14.20
14.03
13.88
13.74
13.61
13.50
13.39
13.29
12.61
12.22
11.97
11.38
500004
999.0
148.5
61.25
37.12
27.00
21.69
18.49
16.39
14.91
13.81
12.97
12.31
11.78
11.34
10.97
10.66
10.39
10.16
9.95
9.77
9.61
9.47
9.34
9.22
9.12
9.02
8.93
8.85
8.77
8.25
7.96
7.77
7.32
540387
999.1
141.1
56.18
33.20
23.70
18.77
15.83
13.90
12.55
11.56
10.80
10.21
9.73
9.34
9.01
8.73
8.49
8.28
8.10
7.94
7.80
7.67
7.55
7.45
7.36
7.27
7.19
7.12
7.05
6.59
6.34
6.17
5.78
562506
999.2
137.1
53.44
31.09
21.92
17.20
14.39
12.56
11.28
10.35
9.63
9.07
8.62
8.25
7.94
7.68
7.46
7.27
7.10
6.95
6.81
6.70
6.59
6.49
6.41
6.33
6.25
6.19
6.12
5.70
5.46
5.31
4.95
576412
999.3
134.6
51.71
29.75
20.80
16.21
13.48
11.71
10.48
9.58
8.89
8.35
7.92
7.57
7.27
7.02
6.81
6.62
6.46
6.32
6.19
6.08
5.98
5.89
5.80
5.73
5.66
5.59
5.53
5.13
4.90
4.76
4.42
585943
999.3
132.9
50.53
28.83
20.03
15.52
12.86
11.13
9.93
9.05
8.38
7.86
7.44
7.09
6.80
6.56
6.35
6.18
6.02
5.88
5.76
5.65
5.55
5.46
5.38
5.31
5.24
5.18
5.12
4.73
4.51
4.37
4.04
592881
999.4
131.6
49.66
28.16
19.46
15.02
12.40
10.70
9.52
8.66
8.00
7.49
7.08
6.74
6.46
6.22
6.02
5.85
5.69
5.56
5.44
5.33
5.23
5.15
5.07
5.00
4.93
4.87
4.82
4.44
4.22
4.09
3.77
598151
999.4
130.6
49.00
27.65
19.03
14.63
12.05
10.37
9.20
8.35
7.71
7.21
6.80
6.47
6.19
5.96
5.76
5.59
5.44
5.31
5.19
5.09
4.99
4.91
4.83
4.76
4.69
4.64
4.58
4.21
4.00
3.86
3.55
602292
999.4
129.9
48.47
27.24
18.69
14.33
11.77
10.11
8.96
8.12
7.48
6.98
6.58
6.26
5.98
5.75
5.56
5.39
5.24
5.11
4.99
4.89
4.80
4.71
4.64
4.57
4.50
4.45
4.39
4.02
3.82
3.69
3.38
605625
999.4
129.3
48.05
26.92
18.41
14.08
11.54
9.89
8.75
7.92
7.29
6.80
6.40
6.08
5.81
5.58
5.39
5.22
5.08
4.95
4.83
4.73
4.64
4.56
4.48
4.41
4.35
4.29
4.24
3.87
3.67
3.54
3.24
610676
999.4
128.3
47.41
26.42
17.99
13.71
11.19
9.57
8.45
7.63
7.00
6.52
6.13
5.81
5.55
5.32
5.13
4.97
4.82
4.70
4.58
4.48
4.39
4.31
4.24
4.17
4.11
4.05
4.00
3.64
3.44
3.32
3.02
615772
999.4
127.4
46.76
25.91
17.56
13.32
10.84
9.24
8.13
7.32
6.71
6.23
5.85
5.54
5.27
5.05
4.87
4.70
4.56
4.44
4.33
4.23
4.14
4.06
3.99
3.92
3.86
3.80
3.75
3.40
3.20
3.08
2.78
620913
999.5
126.4
46.10
25.39
17.12
12.93
10.48
8.90
7.80
7.01
6.40
5.93
5.56
5.25
4.99
4.78
4.59
4.43
4.29
4.17
4.06
3.96
3.87
3.79
3.72
3.66
3.60
3.54
3.49
3.14
2.95
2.83
2.53
623504
999.5
125.9
45.77
25.13
16.90
12.73
10.30
8.72
7.64
6.85
6.25
5.78
5.41
5.10
4.85
4.63
4.45
4.29
4.15
4.03
3.92
3.82
3.74
3.66
3.59
3.52
3.46
3.41
3.36
3.01
2.82
2.69
2.40
626107
999.5
125.5
45.43
24.87
16.67
12.53
10.11
8.55
7.47
6.68
6.09
5.63
5.25
4.95
4.70
4.48
4.30
4.14
4.00
3.88
3.78
3.68
3.59
3.52
3.44
3.38
3.32
3.27
3.22
2.87
2.68
2.55
2.26
628720
999.5
125.0
45.09
24.60
16.44
12.33
9.92
8.37
7.30
6.52
5.93
5.47
5.10
4.80
4.54
4.33
4.15
3.99
3.86
3.74
3.63
3.53
3.45
3.37
3.30
3.23
3.18
3.12
3.07
2.73
2.53
2.41
2.11
631345
999.5
124.5
44.75
24.33
16.21
12.12
9.73
8.19
7.12
6.35
5.76
5.30
4.94
4.64
4.39
4.18
4.00
3.84
3.70
3.58
3.48
3.38
3.29
3.22
3.15
3.08
3.02
2.97
2.92
2.57
2.38
2.25
1.95
633980
999.5
124.0
44.40
24.06
15.98
11.91
9.53
8.00
6.94
6.18
5.59
5.14
4.77
4.47
4.23
4.02
3.84
3.68
3.54
3.42
3.32
3.22
3.14
3.06
2.99
2.92
2.86
2.81
2.76
2.41
2.21
2.08
1.77
2↓
575
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
40
50
60
120
APPENDIX E
CONFIDENCE INTERVALS
FOR THE MEAN
Table E.1 represents 1000 95% confidence intervals constructed from sample
sets selected from a population with a mean () of 25.4 and a variance (2)
of 1.69. These data are discussed in Chapter 4. (Note: Intervals with asterisks
fail to include .)
576
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
TABLE E.1
577
(24.21,
(23.99,
(24.58,
(24.80,
(24.59,
(24.86,
(24.84,
(24.64,
(24.85,
(25.23,
(25.11,
(25.20,
(24.26,
(24.56,
(24.72,
(24.83,
(24.31,
(24.76,
(25.11,
(24.21,
(25.16,
(24.53,
(24.84,
(24.84,
(24.50,
(24.44,
(24.41,
(25.41,
25.94)
25.39)*
26.10)
26.18)
26.09)
25.81)
25.92)
25.93)
26.26)
26.34)
26.18)
26.34)
25.48)
26.17)
25.91)
26.09)
25.89)
25.56)
26.45)
25.86)
26.21)
25.86)
26.31)
26.14)
25.88)
25.85)
25.97)
26.81)*
(24.77,
(25.33,
(24.47,
(24.95,
(24.99,
(24.40,
(24.97,
(24.79,
(25.11,
(24.58,
(24.73,
(24.94,
(24.50,
(24.48,
(24.94,
(24.62,
(24.35,
(25.04,
(24.69,
(25.00,
(24.62,
(24.89,
(24.68,
(25.13,
(24.49,
(24.29,
(25.04,
(24.17,
26.22)
26.45)
25.50)
26.07)
26.15)
25.92)
26.47)
26.18)
26.44)
26.35)
25.83)
25.99)
26.26)
25.76)
26.24)
25.96)
25.73)
26.20)
26.00)
26.52)
26.01)
26.21)
25.55)
26.52)
25.73)
25.29)*
26.23)
25.61)
(24.71,
(24.62,
(25.14,
(24.61,
(24.65,
(24.86,
(25.13,
(24.68,
(24.58,
(24.81,
(24.46,
(24.53,
(25.11,
(24.72,
(24.91,
(25.10,
(24.67,
(24.25,
(24.45,
(25.05,
(24.45,
(24.52,
(24.69,
(25.07,
(25.20,
(24.46,
(24.32,
(25.10,
26.01)
25.89)
26.56)
26.27)
26.29)
25.80)
26.38)
25.93)
25.99)
25.88)
25.51)
25.90)
26.56)
26.02)
26.39)
26.29)
26.07)
26.06)
25.42)
26.55)
25.88)
25.83)
26.28)
26.44)
26.23)
25.94)
25.57)
26.48)
(25.43,
(24.79,
(24.79,
(25.06,
(25.08,
(24.58,
(24.39,
(24.17,
(24.89,
(24.37,
(24.92,
(24.73,
(23.96,
(25.63,
(25.46,
(24.78,
(24.86,
(24.33,
(24.72,
(24.47,
(25.60,
(25.08,
(24.82,
(24.52,
(24.64,
(24.46,
(24.76,
(24.78,
27.01)*
26.16)
25.72)
26.29)
26.28)
25.91)
26.32)
25.83)
26.25)
25.62)
25.87)
26.07)
25.61)
26.40)*
26.82)*
26.24)
26.58)
25.69)
26.50)
25.76)
26.70)*
26.22)
26.02)
25.88)
26.30)
25.45)
25.85)
26.18)
(24.22,
(24.85,
(24.22,
(24.69,
(24.74,
(23.69,
(24.15,
(24.29,
(24.79,
(24.79,
(24.46,
(24.56,
(24.94,
(24.82,
(24.84,
(24.50,
(24.52,
(24.99,
(24.54,
(24.76,
(24.32,
(25.22,
(25.29,
(24.62,
(24.73,
(24.66,
(24.60,
(24.85,
26.19)
26.11)
25.79)
26.25)
26.02)
25.25)*
26.16)
26.14)
26.36)
26.07)
26.45)
25.75)
26.46)
26.09)
26.37)
26.30)
25.55)
26.16)
25.80)
26.07)
25.83)
26.71)
26.44)
26.04)
26.16)
25.78)
26.03)
26.01)
(25.12,
(23.92,
(25.03,
(24.64,
(25.19,
(24.64,
(24.80,
(24.43,
(25.23,
(24.82,
(24.40,
(24.49,
(24.63,
(24.97,
(25.05,
(24.60,
(24.81,
(25.13,
(24.57,
(24.80,
(25.11,
(24.25,
(24.84,
(24.84,
(25.16,
(24.07,
(25.04,
(24.87,
26.56)
25.67)
26.27)
25.88)
26.43)
25.77)
26.01)
25.62)
26.42)
26.19)
25.61)
25.85)
26.31)
26.02)
26.07)
26.20)
26.47)
26.43)
25.76)
25.95)
26.12)
25.39)*
26.21)
25.88)
26.30)
25.62)
26.50)
25.88)
(24.70,
(25.30,
(24.63,
(24.96,
(24.54,
(24.88,
(24.87,
(24.58,
(25.30,
(24.66,
(24.49,
(24.49,
(24.37,
(24.51,
(24.72,
(25.38,
(24.45,
(24.68,
(25.10,
(24.45,
(24.59,
(24.73,
(24.83,
(24.34,
(24.66,
(24.48,
(24.39,
(24.62,
26.54)
26.84)
25.83)
26.15)
25.73)
26.12)
26.54)
25.99)
26.45)
26.56)
26.45)
26.24)
25.95)
25.78)
25.78)
26.62)
26.32)
26.26)
26.16)
25.89)
25.81)
26.34)
26.03)
25.71)
25.87)
25.88)
26.27)
26.00)
(24.50, 25.95)
(24.71, 26.39)
(25.09, 26.84)
(25.04, 25.92)
(24.17, 25.94)
(24.63, 25.92)
(24.66, 25.95)
(25.11, 26.33)
(25.09, 26.53)
(24.56, 26.12)
(24.61, 26.02)
(25.28, 26.26)
(24.67, 26.18)
(24.88, 26.18)
(25.20, 26.60)
(24.86, 26.20)
(23.94, 25.77)
(24.69, 25.68)
(24.60, 25.90)
(24.82, 26.86)
(24.11, 25.86)
(24.67, 26.24)
(24.90, 26.26)
(24.25, 25.51)
(25.20, 26.59)
(24.99, 26.61)
(24.97, 26.13)
(25.21, 26.39)
(continues)
578
TABLE E.1 (Continued )
(24.98,
(24.83,
(25.07,
(24.61,
(24.41,
(24.36,
(24.56,
(25.19,
(24.23,
(25.17,
(24.42,
(25.43,
(24.85,
(24.24,
(24.75,
(24.42,
(24.50,
(25.11,
(25.06,
(24.53,
(24.56,
(24.82,
(25.05,
(24.76,
(24.52,
(24.74,
(24.71,
(24.55,
26.33)
26.27)
26.38)
26.21)
25.51)
25.85)
25.98)
26.85)
26.01)
26.78)
25.86)
26.53)*
26.42)
26.00)
26.11)
25.94)
25.73)
26.24)
26.28)
26.19)
26.21)
25.75)
26.31)
26.37)
26.21)
25.92)
26.55)
25.94)
(24.11,
(25.07,
(24.90,
(24.93,
(24.93,
(24.45,
(25.19,
(24.38,
(24.28,
(24.52,
(24.61,
(25.00,
(25.49,
(24.00,
(24.82,
(24.71,
(24.96,
(24.14,
(24.26,
(24.77,
(24.44,
(24.68,
(24.89,
(25.30,
(24.85,
(24.45,
(25.01,
(24.79,
26.09)
26.29)
26.44)
26.12)
26.11)
25.68)
26.24)
25.92)
25.56)
26.20)
26.04)
26.32)
26.78)*
25.11)*
26.07)
26.39)
26.19)
25.70)
25.69)
26.25)
25.76)
25.98)
26.20)
26.54)
25.86)
25.68)
26.44)
26.08)
(24.74,
(25.24,
(24.44,
(24.41,
(24.90,
(24.84,
(24.27,
(24.59,
(24.14,
(24.29,
(25.13,
(24.86,
(25.52,
(24.85,
(24.62,
(24.15,
(25.35,
(24.94,
(24.30,
(25.15,
(24.27,
(25.01,
(24.12,
(25.05,
(25.16,
(24.89,
(24.67,
(24.94,
26.05)
26.56)
25.72)
25.79)
25.91)
25.91)
25.82)
26.19)
25.37)*
25.86)
26.18)
25.70)
26.38)*
26.03)
25.84)
25.84)
26.41)
25.85)
25.42)
26.49)
25.50)
26.78)
25.66)
26.20)
25.95)
26.23)
26.20)
25.90)
(24.11,
(24.13,
(24.63,
(24.95,
(25.18,
(24.83,
(24.98,
(24.76,
(24.64,
(24.26,
(24.09,
(24.52,
(24.14,
(24.73,
(25.42,
(24.72,
(24.58,
(24.59,
(25.03,
(24.96,
(25.09,
(24.83,
(24.79,
(24.38,
(24.71,
(24.80,
(24.94,
(24.73,
25.65)
25.65)
25.95)
26.06)
26.44)
26.40)
26.00)
26.42)
26.29)
25.97)
25.73)
25.65)
25.61)
26.02)
26.48)*
25.98)
25.92)
25.96)
26.56)
26.12)
26.03)
26.36)
26.43)
25.72)
25.92)
26.04)
26.35)
25.87)
(24.47,
(25.41,
(25.28,
(24.92,
(24.90,
(24.69,
(24.85,
(24.56,
(24.96,
(24.74,
(24.83,
(24.67,
(25.03,
(25.19,
(24.75,
(25.22,
(24.23,
(25.02,
(24.47,
(25.13,
(24.76,
(25.39,
(24.98,
(24.89,
(24.65,
(24.84,
(24.79,
(24.40,
25.83)
26.47)*
26.37)
25.88)
26.48)
26.04)
25.98)
26.00)
26.40)
26.17)
26.19)
26.00)
26.43)
26.92)
26.03)
26.51)
25.64)
26.60)
26.10)
26.32)
26.01)
26.46)
26.15)
26.37)
25.36)*
26.24)
26.36)
25.68)
(24.85,
(24.64,
(24.63,
(24.69,
(25.06,
(24.68,
(25.00,
(24.57,
(24.89,
(24.81,
(25.18,
(25.17,
(24.38,
(24.70,
(24.32,
(24.16,
(25.18,
(24.67,
(24.52,
(24.69,
(24.63,
(24.39,
(25.05,
(24.34,
(25.00,
(24.15,
(24.98,
(25.24,
26.32)
26.04)
25.96)
26.10)
26.12)
26.11)
25.82)
26.29)
26.33)
26.20)
26.53)
26.69)
26.29)
26.18)
25.70)
25.76)
26.28)
26.26)
25.77)
25.99)
26.14)
25.52)
26.57)
25.68)
26.48)
25.62)
26.31)
26.39)
(24.37,
(24.42,
(24.63,
(24.84,
(25.23,
(24.95,
(24.72,
(24.60,
(24.43,
(24.97,
(25.01,
(24.73,
(25.16,
(24.85,
(25.19,
(25.18,
(24.36,
(24.83,
(25.12,
(24.63,
(24.36,
(25.07,
(24.81,
(24.64,
(24.58,
(25.48,
(24.61,
(25.18,
25.63)
25.94)
25.71)
26.22)
26.18)
26.17)
25.99)
26.12)
25.11)*
26.15)
26.66)
25.98)
26.73)
26.40)
26.45)
25.94)
26.29)
26.38)
26.44)
26.15)
25.59)
26.11)
26.08)
25.75)
25.61)
27.00)*
26.21)
26.61)
(24.66,
(24.61,
(24.98,
(25.40,
(24.87,
(24.94,
(24.98,
(24.99,
(24.88,
(24.85,
(24.52,
(24.58,
(24.92,
(24.84,
(24.71,
(25.04,
(24.18,
(24.43,
(25.55,
(25.37,
(24.36,
(24.97,
(24.32,
(24.54,
(24.66,
(25.00,
(23.94,
(24.42,
25.95)
25.61)
26.38)
26.67)
26.06)
26.00)
26.48)
26.18)
26.53)
26.43)
25.97)
25.93)
26.35)
26.31)
26.31)
26.87)
25.74)
25.38)*
26.84)*
26.48)
26.04)
26.26)
25.69)
26.08)
26.26)
26.34)
25.88)
25.61)
(24.36,
(25.57,
(25.46,
(24.33,
(25.00,
(24.95,
(25.14,
(24.60,
(24.97,
(24.18,
(24.67,
(24.30,
(24.39,
(24.04,
(24.89,
(24.76,
(25.08,
(24.52,
(24.97,
(24.54,
(24.74,
(24.34,
(24.37,
(24.58,
(24.44,
(24.72,
(24.27,
(25.09,
25.77)
26.72)*
26.24)*
26.14)
25.95)
26.22)
26.48)
25.65)
26.37)
25.96)
25.80)
25.75)
25.48)
25.44)
26.28)
26.25)
26.29)
25.99)
26.44)
25.79)
26.67)
25.83)
25.61)
25.81)
26.16)
26.08)
25.94)
26.67)
(24.35,
(24.85,
(25.25,
(25.00,
(24.51,
(24.94,
(24.63,
(24.75,
(24.42,
(24.59,
(24.59,
(25.33,
(24.89,
(24.83,
(24.05,
(24.37,
(24.60,
(24.89,
(24.59,
(25.13,
(25.03,
(25.10,
(24.49,
(24.47,
(24.70,
(24.73,
(24.93,
(24.44,
25.70)
26.11)
26.39)
26.19)
26.03)
26.28)
26.28)
26.31)
26.73)
25.61)
26.10)
26.18)
26.44)
26.22)
25.43)
25.53)
26.08)
26.14)
25.63)
26.68)
26.54)
26.12)
25.97)
25.66)
26.25)
26.17)
26.73)
25.88)
(24.85,
(24.98,
(24.70,
(23.88,
(24.24,
(24.75,
(24.91,
(24.23,
(24.63,
(24.73,
(23.87,
(24.18,
(24.79,
(24.43,
(24.78,
(24.85,
(25.24,
(24.29,
(24.80,
(24.60,
(24.69,
(24.87,
(24.48,
(24.39,
(24.48,
(25.44,
(24.43,
(24.59,
26.06)
26.41)
26.22)
25.22)*
25.97)
26.00)
25.91)
26.01)
25.53)
25.86)
25.38)*
26.00)
26.14)
25.72)
26.06)
25.86)
26.50)
25.84)
26.00)
25.78)
26.59)
26.48)
25.60)
25.72)
25.68)
26.34)*
25.84)
26.63)
(24.56,
(24.71,
(24.59,
(24.42,
(24.92,
(24.78,
(25.08,
(24.94,
(24.48,
(24.63,
(24.93,
(24.58,
(24.62,
(24.38,
(24.50,
(24.26,
(24.79,
(24.68,
(24.91,
(24.49,
(24.84,
(25.00,
(24.65,
(24.12,
(24.32,
(24.43,
(24.29,
(24.92,
25.51)
26.00)
25.88)
26.05)
26.35)
26.37)
26.23)
26.36)
25.73)
25.97)
26.46)
26.18)
26.04)
26.20)
25.94)
25.48)
26.10)
25.99)
26.29)
25.87)
26.37)
26.54)
26.73)
25.71)
25.64)
25.75)
25.52)
26.37)
(24.23,
(25.26,
(24.43,
(24.68,
(24.45,
(25.45,
(24.47,
(24.66,
(24.44,
(24.09,
(24.69,
(25.04,
(25.29,
(24.63,
(24.54,
(25.14,
(25.17,
(24.41,
(24.56,
(24.94,
(24.81,
(24.64,
(25.18,
(24.58,
(24.65,
(25.27,
(24.55,
(25.10,
25.69)
26.38)
25.54)
26.02)
26.33)
27.04)*
25.85)
26.01)
25.66)
25.90)
26.16)
26.25)
26.58)
25.63)
26.01)
26.41)
26.53)
26.17)
25.92)
26.19)
26.35)
26.00)
26.41)
25.85)
25.50)
26.74)
26.35)
26.29)
(24.41,
(24.71,
(25.09,
(24.93,
(24.94,
(24.20,
(24.54,
(23.84,
(24.14,
(25.47,
(24.66,
(24.33,
(24.81,
(24.40,
(24.97,
(24.17,
(24.85,
(25.18,
(24.65,
(25.13,
(24.64,
(24.71,
(25.09,
(25.77,
(24.75,
(24.13,
(25.13,
(24.57,
25.71)
26.36)
26.50)
26.18)
26.08)
25.91)
26.04)
25.91)
25.11)*
26.45)*
25.94)
26.15)
26.29)
26.17)
25.89)
25.84)
25.88)
26.57)
26.02)
26.39)
25.95)
26.09)
26.44)
27.28)*
26.21)
25.49)
26.54)
26.27)
(24.52,
(24.66,
(24.83,
(24.85,
(24.07,
(24.71,
(24.95,
(24.78,
(24.84,
(24.69,
(24.48,
(24.58,
(24.35,
(25.09,
(24.58,
(24.77,
(24.79,
(24.10,
(24.34,
(24.47,
(24.89,
(24.43,
(24.95,
(24.67,
(24.25,
(25.04,
(24.27,
(24.49,
25.89)
25.99)
26.22)
26.20)
26.38)
26.08)
26.74)
26.42)
26.33)
25.92)
25.95)
26.13)
25.28)*
26.50)
25.87)
26.03)
26.18)
25.54)
25.42)
25.60)
26.07)
26.05)
26.40)
25.86)
25.69)
26.28)
25.63)
25.90)
579
(24.53, 25.55)
(24.89, 25.86)
(24.87, 26.11)
(25.52, 26.77)*
(24.84, 26.27)
(24.14, 25.83)
(25.00, 26.40)
(25.13, 26.34)
(25.05, 26.14)
(24.58, 25.77)
(24.33, 26.04)
(24.32, 25.70)
(25.32, 26.60)
(24.93, 26.45)
(24.86, 26.11)
(25.26, 26.53)
(24.64, 26.11)
(25.23, 26.47)
(24.86, 26.32)
(24.61, 25.77)
(24.50, 25.91)
(24.54, 25.69)
(24.64, 26.01)
(24.78, 26.43)
(24.10, 25.83)
(24.63, 25.80)
(25.17, 26.45)
(24.76, 26.08)
(continues)
580
TABLE E.1 (Continued )
(24.64,
(24.44,
(25.53,
(25.72,
(24.45,
(24.95,
(24.83,
(24.75,
(24.69,
(24.88,
(24.32,
(24.55,
(25.18,
(24.60,
(24.69,
(24.89,
(24.95,
(24.84,
(24.65,
(24.86,
(24.65,
(24.73,
(25.14,
(24.27,
(24.70,
(24.60,
(24.20,
(24.62,
26.13)
26.00)
26.94)*
26.82)*
26.06)
26.13)
26.42)
26.15)
25.95)
26.52)
25.93)
25.95)
26.27)
26.16)
25.83)
26.16)
26.36)
25.95)
26.09)
26.81)
26.09)
26.32)
26.47)
25.36)*
26.04)
26.25)
25.79)
25.87)
(24.46,
(24.46,
(23.89,
(24.36,
(25.19,
(25.17,
(24.62,
(25.33,
(24.74,
(25.04,
(24.74,
(24.33,
(24.10,
(24.83,
(25.00,
(24.99,
(24.37,
(25.06,
(24.96,
(24.22,
(25.23,
(24.42,
(24.92,
(24.37,
(24.88,
(25.52,
(24.52,
(24.54,
25.84)
25.63)
25.24)*
25.75)
26.42)
26.46)
26.12)
26.63)
26.22)
26.48)
26.00)
25.56)
25.13)*
26.07)
26.41)
26.73)
25.92)
26.51)
26.36)
25.71)
26.49)
25.81)
26.19)
25.56)
26.14)
26.51)*
25.70)
26.05)
(24.31,
(24.75,
(24.78,
(24.63,
(24.64,
(24.70,
(24.39,
(24.69,
(24.61,
(24.71,
(24.42,
(24.22,
(24.64,
(25.63,
(25.06,
(24.85,
(24.81,
(25.09,
(24.88,
(24.98,
(25.25,
(24.63,
(24.48,
(25.35,
(24.45,
(24.68,
(25.58,
(24.07,
25.89)
26.33)
26.12)
26.25)
26.18)
26.19)
25.41)
25.90)
26.15)
25.90)
25.72)
25.61)
25.69)
26.64)*
26.19)
26.24)
26.10)
25.96)
26.18)
26.46)
26.83)
25.70)
25.91)
26.84)
25.86)
26.20)
26.83)*
25.80)
(25.34,
(25.03,
(24.95,
(24.65,
(24.95,
(24.59,
(25.18,
(24.67,
(25.10,
(24.86,
(25.14,
(24.59,
(25.16,
(24.77,
(24.32,
(24.91,
(24.89,
(25.44,
(24.94,
(24.29,
(24.97,
(24.66,
(25.13,
(24.73,
(24.30,
(24.87,
(25.15,
(25.01,
26.66)
26.03)
26.19)
26.19)
26.25)
25.94)
26.22)
26.29)
26.15)
26.18)
26.53)
26.25)
26.15)
26.01)
26.34)
26.12)
26.40)
26.34)*
26.05)
25.43)
25.94)
25.43)
26.42)
26.10)
26.08)
26.18)
25.94)
26.40)
(24.87,
(24.87,
(24.81,
(25.04,
(24.43,
(24.71,
(24.75,
(24.89,
(24.95,
(24.45,
(24.74,
(25.21,
(24.95,
(24.41,
(25.11,
(24.79,
(24.80,
(25.09,
(24.70,
(24.53,
(24.88,
(25.58,
(24.15,
(24.80,
(24.66,
(24.57,
(24.87,
(25.08,
26.31)
26.36)
26.05)
25.80)
26.35)
26.06)
26.33)
26.24)
26.40)
25.96)
25.97)
26.31)
26.03)
25.84)
26.42)
26.03)
25.90)
26.27)
25.78)
26.00)
26.13)
26.41)*
25.45)
26.20)
25.43)
25.65)
26.34)
26.52)
(24.17,
(24.45,
(24.76,
(25.01,
(24.89,
(24.84,
(24.47,
(24.35,
(24.46,
(24.82,
(24.35,
(25.03,
(24.91,
(25.38,
(23.99,
(24.52,
(24.60,
(25.19,
(25.10,
(24.58,
(24.37,
(24.22,
(24.84,
(24.35,
(24.46,
(25.60,
(24.70,
(24.15,
25.47)
25.83)
26.13)
26.66)
26.11)
26.45)
26.37)
26.27)
26.08)
25.94)
26.08)
26.40)
26.08)
26.21)
25.83)
25.83)
26.12)
26.86)
26.22)
26.31)
25.41)
26.04)
26.14)
25.81)
26.12)
26.77)*
26.12)
25.80)
(25.28,
(24.85,
(24.38,
(24.48,
(24.19,
(24.41,
(24.45,
(24.95,
(25.07,
(24.61,
(24.70,
(25.09,
(24.83,
(24.53,
(24.21,
(24.74,
(24.84,
(24.97,
(24.63,
(24.27,
(25.37,
(24.43,
(24.77,
(24.30,
(24.38,
(24.91,
(24.40,
(24.59,
26.57)
26.15)
25.77)
25.96)
25.61)
25.60)
25.74)
26.01)
26.68)
26.09)
25.75)
26.54)
25.95)
25.86)
25.43)
26.21)
26.32)
26.15)
26.06)
25.58)
26.31)
26.08)
26.08)
25.74)
25.55)
26.44)
25.56)
26.22)
(24.75,
(24.12,
(24.45,
(24.88,
(24.52,
(24.76,
(25.19,
(24.77,
(25.45,
(24.87,
(25.11,
(24.58,
(24.49,
(24.17,
(24.84,
(24.37,
(25.02,
(25.31,
(24.77,
(25.25,
(24.80,
(24.81,
(25.22,
(24.15,
(25.13,
(24.88,
(24.93,
(24.65,
26.14)
25.75)
26.24)
26.48)
26.26)
25.96)
26.37)
25.94)
26.28)*
26.67)
26.29)
25.72)
25.69)
25.80)
25.99)
26.01)
26.19)
26.21)
26.36)
26.68)
26.43)
26.14)
26.36)
25.64)
26.01)
26.14)
26.21)
25.90)
(25.21,
(24.40,
(24.79,
(24.22,
(24.44,
(24.46,
(24.92,
(25.02,
(24.98,
(25.13,
(24.80,
(25.34,
(25.38,
26.21)
25.75)
25.71)
25.61)
25.73)
25.86)
25.76)
26.02)
26.21)
26.29)
26.15)
26.62)
26.61)
(24.85,
(25.29,
(23.90,
(24.39,
(24.75,
(24.92,
(24.65,
(25.01,
(24.78,
(24.36,
(24.77,
(24.29,
(25.28,
26.46)
26.31)
25.01)*
26.25)
26.44)
26.30)
26.04)
26.38)
26.11)
25.74)
26.08)
25.38)*
26.67)
(25.22,
(24.51,
(24.61,
(24.81,
(25.03,
(24.65,
(24.96,
(24.71,
(24.82,
(24.38,
(24.52,
(24.89,
(24.76,
26.74)
26.07)
26.39)
26.08)
25.84)
26.25)
26.33)
26.19)
26.44)
25.42)
26.03)
26.23)
26.06)
(25.07,
(24.80,
(24.89,
(25.41,
(24.53,
(24.85,
(24.97,
(24.45,
(24.75,
(24.83,
(24.42,
(24.89,
(25.10,
26.45)
26.23)
26.53)
26.35)*
26.10)
26.22)
26.55)
25.26)*
26.34)
26.40)
25.97)
26.19)
26.19)
(24.05,
(24.82,
(24.86,
(24.62,
(25.05,
(24.22,
(25.30,
(24.55,
(25.21,
(24.23,
(25.47,
(24.85,
(24.65,
25.59)
26.58)
26.49)
25.97)
26.12)
25.46)
26.55)
26.16)
26.27)
25.42)
26.93)*
26.08)
25.90)
(24.08,
(24.81,
(24.54,
(25.02,
(25.21,
(25.24,
(25.15,
(24.90,
(24.55,
(24.34,
(24.56,
(24.84,
(24.55,
25.41)
26.06)
25.86)
26.13)
26.39)
26.42)
26.62)
26.09)
25.71)
25.89)
25.93)
26.24)
26.19)
(24.24,
(24.74,
(24.54,
(24.39,
(24.37,
(24.63,
(24.82,
(24.71,
(24.94,
(24.29,
(24.53,
(24.89,
(24.96,
25.36)*
26.04)
26.25)
25.76)
25.81)
25.97)
25.93)
26.29)
26.42)
26.05)
25.74)
26.15)
26.05)
(24.67,
(24.67,
(24.67,
(24.53,
(25.16,
(24.43,
(25.05,
(25.14,
(24.34,
(24.01,
(24.72,
(24.60,
(24.48,
26.08)
25.76)
25.82)
25.72)
26.25)
25.65)
26.51)
26.43)
25.64)
25.61)
26.14)
25.76)
25.98)
581
APPENDIX F
MAP PROJECTION
COORDINATE SYSTEMS
F.1
INTRODUCTION
Most local surveyors are well served by using map projections such as the
state plane coordinate system. These two-dimensional grid systems allow surveyors to perform accurate computations over large regions of land using
plane surveying computations. They are the basis for the adjustments discussed in Chapters 14 through 16.
Map projections provide a one-to-one mathematical relationship with
points on the ellipsoid and those on the mapping surface. There are an infinite
number of map projections. Most map projections are defined by a series of
mathematical transformations used to convert a point’s geodetic coordinates
of latitude, , and longitude, , to xy grid coordinates. Some map projections
preserve the shape of objects (conformal); others, areas, directions, or distances of lines. However, since Earth is ellipsoidal in shape and a mapping
surface is a plane, all map projections introduce some form of distortion to
observations. For example, distances and areas are distorted in a conformal
map projection.
To reduce the size of these distortions, the developable surface is often
made secant to the ellipsoid and the width of the mapping zone is limited in
distance. For instance, when the National Geodetic Survey originally designed
the state plane coordinate system during the 1930s, the zone widths were
limited to 158 miles so that precision between the ellipsoid distance and the
grid distance was no worse than 1⬊10,000. Since most surveys at that time
were only accurate to a precision of 1⬊5000, this was an acceptable limit.
However, with today’s modern instruments, observations must be reduced
properly if survey accuracy is to be preserved in a map projection system.
582
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
F.2
MATHEMATICS OF THE LAMBERT CONFORMAL CONIC MAP PROJECTION
583
All map projections are based on the ellipsoid selected, such as the Geodetic Reference System of 1980 (GRS 80), and defining zone parameters.
Typically, the zone parameters define the grid origin (0, 0); the secant lines
of the projection, also known as standard parallels, or scale factor, k0, at the
central meridian, 0; and the offset distances (E0, Nb) from the grid origin.
Once defined, each map projection has a series of zone constants that are
computed using the defining zone parameters. These zone constants are computed only once for each projection. Once the zone constants are computed,
the direct and inverse problems can be carried out for any point in the system.
The direct problem takes the geodetic coordinates of a point and transforms
them into grid coordinates, and the inverse problem takes the grid coordinates
of a point and transforms them into geodetic coordinates.
The two primary map projection systems used in the United States are the
Lambert Conformal Conic for states that have a long east–west extent and
the Transverse Mercator for states that have a long north–south extent. Both
map projections are conformal; that is, they preserve angles in infinitesimally
small regions about a point. This property is advantageous to surveyors since
angles are minimally distorted when using a conformal projection. On the
other hand, as shown in Figure F.1, horizontal distances observed must be
reduced to the mapping surface to eliminate the distortions of the projection.
However, if these reductions are performed properly, the resulting plane computations are as accurate as geodetic computations such as those shown in
Chapter 23. In this appendix we look at the mathematics of the Lambert
Conformal Conic and Transverse Mercator map projections and demonstrate
proper methods in reducing observations before an adjustment.
F.2 MATHEMATICS OF THE LAMBERT CONFORMAL
CONIC MAP PROJECTION
The Lambert Conformal Conic map projection was introduced by Johann
Lambert in 1772. As its name implies, this map projection uses a cone as its
developable surface. The projection is conformal, so angles are preserved but
distances are distorted. A Lambert Conformal Conic map projection is defined
by two ellipsoidal parameters,1 grid origin (0, 0); latitude of the north standard parallel, N, and south standard parallel,2 S; false easting, E0; and false
northing, Nb.
1
Typically, an ellipsoid is defined by the length of its semimajor axis, a, and its flattening factor,
ƒ. The first eccentricity is computed as e ⫽ 2ƒ ⫺ ƒ2. The GRS 80 ellipsoid has defining
parameters of a ⫽ 6,378,137.0 m and ƒ ⫽ 1 / 298.2572221008.
2
The standard parallels are the latitudes of the north and south secant lines for the cone on the
ellipsoid.
584
MAP PROJECTION COORDINATE SYSTEMS
Lm
H
N
Geoid
Ellipsoid
Mapping surface
Re
O
Figure F.1 Reduction of distance to a mapping surface.
F.2.1
Zone Constants
A set of three functions is used repeatedly in computations of the Lambert
Conformal Conic map projection:
W() ⫽ 1 ⫺ e2 sin2
(F.1)
M() ⫽
cos
W()
(F.2)
T() ⫽
冪
冉
冊
1 ⫺ sin 1 ⫹ e sin
1 ⫹ sin 1 ⫺ e sin
e
(F.3)
Using Equations (F.1) through (F.3), the remaining zone constants are defined as
w1 ⫽ W(S)
(F.4)
w2 ⫽ W(N)
(F.5)
m1 ⫽ M(S)
(F.6)
m2 ⫽ M(N)
(F.7)
t0 ⫽ T(0)
(F.8)
t1 ⫽ T(S)
(F.9)
F.2
MATHEMATICS OF THE LAMBERT CONFORMAL CONIC MAP PROJECTION
t2 ⫽ T(N)
n ⫽ sin 0 ⫽
F⫽
(F.10)
ln m1 ⫺ ln m2
ln t1 ⫺ ln t2
m1
nt n1
Rb ⫽ aFt 0n ⫽ radius of the projection
F.2.2
585
(F.11)
(F.12)
(F.13)
Direct Problem
The direct problem takes the geodetic coordinates of latitude, , and longitude, , of a point and transforms them into xy grid coordinates. Often, the y
coordinate is called the point’s northing, N, and the x coordinate its easting,
E. Thus, given the geodetic coordinates of a point, the northing, y, easting,
x, scale factor, k, and convergence angle, ␥, of the point are computed as
t ⫽ T()
(F.14)
m ⫽ M()
(F.15)
R ⫽ aFt n
(F.16)
␥ ⫽ ( ⫺ 0)n
(where western longitude is considered negative)
(F.17)
E ⫽ R sin ␥ ⫹ E0
(F.18)
N ⫽ Rb ⫺ R cos ␥ ⫹ Nb
(F.19)
k⫽
F.2.3
Rn
am
(F.20)
Inverse Problem
The inverse problem takes a point’s northing and easting coordinates and
computes its latitude, longitude, scale factor, and convergence angle. For the
Lambert Conformal Conic map projection, the equations for the inverse problem are
586
MAP PROJECTION COORDINATE SYSTEMS
E⬘ ⫽ E ⫺ E0
(F.21)
N⬘ ⫽ Rb ⫺ (N ⫺ Nb)
(F.22)
R ⫽ E⬘2 ⫹ N⬘2
t⫽
(F.23)
冉冊
R
aF
␥ ⫽ tan⫺1
1/n
(F.24)
E⬘
N⬘
(F.25)
⫽ 90⬚ ⫺ tan⫺1t
⫽ 90⬚ ⫺ 2 tan⫺1
(F.26)
冋冉
t
冊册
1 ⫺ e sin
1 ⫹ sin
e/2
(F.27)
Repeat Equation (F.27) using for in the first iteration. Iterate until the
change in is insignificant; that is, the change should be less than 0.000005ⴖ.
⫽
k⫽
␥
n
⫹ 0
m1t n
mt n1
(F.28)
(F.29)
where m and t are defined in Equations (F.14) and (F.15) using from Equation (F.27).
F.3
MATHEMATICS OF THE TRANSVERSE MERCATOR
The Transverse Mercator map projection uses a cylinder as its developable
surface. It preserves scale in a north–south direction and thus is good for
regions with a long north–south extent. This projection was proposed by
Johann Lambert, but the mathematics for an ellipsoid were not solved until
the early twentieth century. In many countries, this projection is also known
as the Gauss–Krüger map projection. The most famous Transverse Mercator
map projection is the Universal Transverse Mercator (UTM) developed by
the National Geospatial-Information Agency to provide a worldwide mapping
system from 80⬚ south latitude to 80⬚ north latitude. This map projection is
F.3
MATHEMATICS OF THE TRANSVERSE MERCATOR
587
defined by two ellipsoidal parameters,3 grid origin (0,0), scale factor, k0, at
the central meridian, 0, false easting, E0, and false northing, Nb.
There are 60 zones in the Universal Transverse Mercator map projection,
each nominally 6⬚ wide. Each zone overlaps its neighboring zones by 30⬘.
The central meridian, 0, for each zone is assigned a false easting, E0, of
500,000 m. The false northing, Nb, is 0.000 m in the northern hemisphere
and 10,000,000.000 m in the southern hemisphere. The scale factor at the
central meridian, k0, is 0.9996, which yields a distance precision of 1⬊2500.
The central meridians (0) for each zone start at 177⬚ west longitude and with
a few exceptions, proceeds easterly by 6⬚ for each subsequent zone. The grid
origins are at 0⬚ and 0.
F.3.1
Zone Constants
The Transverse Mercator map projection use the following defining functions:
C() ⫽ e⬘2 cos2
(F.30)
T() ⫽ tan
(F.31)
冋冉
冉
M() ⫽ a
⫹
1⫺
冊 冉
冊
冉 冊
冊
e2 3e4
5e6
3e2 3e4
45e6
⫺
⫺
⫺
⫹
⫹
sin 2
4
64
256
8
32
1024
册
15e4
45e6
35e6
⫹
sin 4 ⫺
sin 6
256
1024
3072
(F.32)
where e is the first eccentricity of the ellipse as defined in Equation (17.5)
and e⬘ is defined as
b ⫽ a(1 ⫺ ƒ)
e⬘ ⫽
(F.33)
a ⫺ b
2
b
2
冪1 ⫺e e
2
⫽
m0 ⫽ M(0)
3
2
(F.34)
The Universal Transverse Mercator (UTM) uses the WGS 84 ellipsoid, defined in Chapter 17.
588
MAP PROJECTION COORDINATE SYSTEMS
F.3.2
Direct Problem
The equations in the Transverse Mercator for the direct problem are
m ⫽ M()
(F.35)
t ⫽ T()
(F.36)
c ⫽ C()
(F.37)
A ⫽ ( ⫺ 0) cos where western longitudes are negative
(F.38)
E ⫽ k0RN A ⫹ (1 ⫺ t ⫹ c)
A3
6
⫹ (5 ⫺ 18t ⫹ t2 ⫹ 72c ⫺ 58e⬘2)
再
A5
⫹ E0
120
(F.39)
冋
A2
A4
⫹ (5 ⫺ t ⫹ 9c ⫹ 4c2)
2
24
N ⫽ k0 m ⫺ m0 ⫹ RN tan
⫹ (61 ⫺ 58t ⫹ t2 ⫹ 600c ⫺ 300e⬘2)
A6
720
册冎
⫹ Nb
(F.40)
where RN is the radius in the prime vertical as defined by N in Equation
(17.6).
c2 ⫽
1 ⫹ 3c ⫹ 2c2
3
c3 ⫽
2 ⫺ tan2
15
(F.41)
␥ ⫽ A tan [1 ⫹ A2(c2 ⫹ c3A2)]
冋
k ⫽ k0 1 ⫹ (1 ⫹ c)
A2
A4
⫹ (5 ⫺ 4t ⫹ 42c ⫹ 13c2 ⫺ 23e⬘2)
2
24
⫹ (61 ⫺ 148t ⫹ 16t2)
F.3.3
册
A6
720
Inverse Problem
The equations in the Transverse Mercator for the inverse problem are
(F.42)
F.3
MATHEMATICS OF THE TRANSVERSE MERCATOR
589
E⬘ ⫽ E ⫺ E0
(F.43)
N⬘ ⫽ N ⫺ Nb
(F.44)
e1 ⫽
1 ⫺ 兹1 ⫺ e2
(F.45)
1 ⫹ 兹1 ⫺ e2
m ⫽ m0 ⫹
⫽
N⬘
k0
冉
(F.46)
m
2
4
6
冊
(F.47)
a 1 ⫺ e /4 ⫺ 3e /64 ⫺ 5e /256
The foot-point latitude is
ƒ ⫽ ⫹
⫹
冉
冊
冉
冊
3e1 27e31
21e21 55e41
⫺
sin 2 ⫹
⫺
sin 4
2
32
16
32
151e31
1097e41
sin 6 ⫹
sin 8
96
512
(F.48)
Using the foot-point latitude and functions defined in Section F.3.1 and
Equation (23.16) yields
c1 ⫽ C(ƒ)
(F.49)
t1 ⫽ T(ƒ)
(F.50)
N1 ⫽
a
兹1 ⫺ e2 sin2 ƒ
(F.51)
M1 ⫽
a(1 ⫺ e2)
(1 ⫺ e2 sin2 ƒ)3 / 2
(F.52)
D⫽
E⬘
N1k0
(F.53)
B⫽
D2
D4
⫺ (5 ⫹ 3t1 ⫹ 10c1 ⫺ 4c21 ⫺ 9e⬘)
2
24
⫹ (61 ⫹ 90t1 ⫹ 298c1 ⫹ 454t12 ⫺ 252e⬘ ⫺ 3c12)
D6
720
(F.54)
590
MAP PROJECTION COORDINATE SYSTEMS
⫽ ƒ ⫺
N1 tan ƒ
B
M1
(F.55)
D ⫺ (1 ⫺ 2t1 ⫹ c1) (D3 /6)
⫹ (5 ⫺ 2c1 ⫹ 28t1 ⫺ 3c12 ⫹ 8e⬘2 ⫹ 24t12) (D5 /120)
⫽ 0 ⫹
cos ƒ
(F.56)
Note that Equations (F.41) and (F.42) can be used to compute the convergence angle ␥ and scale factor k for the point.
F.4
REDUCTION OF OBSERVATIONS
Most often, the grid coordinates of a point are known prior to the survey and
all that is needed is to reduce the observations to the mapping surface. The
basic principle to bear in mind is that grid computations should only be
performed with grid observations. Since the two map projections discussed
previously are conformal, observed distances must be reduced to the mapping
surface. Similarly, geodetic and astronomical directions must be converted to
their grid equivalents.
As discussed in this section, conformality implies that the angles will be
only slightly distorted. As will be shown, the arc-to-chord correction is applied directions and angles when the sight distances are long. For example,
in the state plane coordinate system, this correction should be considered for
angles whose sight distances are greater than 8 km. In this section, proper
reduction of distance, direction, and angle observations is discussed.
F.4.1
Reduction of Distances
As shown in Figure F.1, an observed horizontal distance must be reduced to
the mapping surface. This reduction usually involves using the grid factor.
The grid factor is the product of the elevation factor, which reduces the observed distance to the ellipsoid, and a scale factor (k), which reduces the
ellipsoidal distance to the mapping surface.
There are several procedures for reducing an observed distance to the ellipsoid, the most precise being a geodetic reduction. However, surveyed
lengths typically contain only five or six significant figures. Thus, less strict
methods can be applied to these short lengths. The elevation factor is computed as
EF ⫽
Re
Re
⫽
Re ⫹ H ⫹ N Re ⫹ h
(F.57)
F.4
REDUCTION OF OBSERVATIONS
591
In Equation (F.57), Re is the radius of the Earth, H the orthometric height,
N the geoidal height, and h the geodetic height. All of these parameters are
determined at the observation station. The relationship between the geodetic
height, h, and orthometric height, H, is
h⫽H⫹N
(F.58)
In Equation (F.57), the radius in the azimuth of the line should be used
for Re. Again since surveyors observe short distances typically, an average
radius of the Earth of 6,371,000 m can be used in computing EF. These
approximations are demonstrated in Example F.1.
In a map projection system, the scale factor computed using Equation
(F.20), (F.29), or (F.42) is for a point. Generally, the scale factor changes
continuously along the length of the line. Thus, a weighted mean using two
endpoints of the line (k1 and k2) and midpoint (km) is a logical choice for
computing a single scale factor for a line. It can be computed as
kavg ⫽
k1 ⫹ 4km ⫹ k2
6
(F.59)
However, as with the elevation factor, this type of precision is seldom needed
for the typical survey. Thus, the mean of the two endpoint scale factors is
generally of sufficient accuracy for most surveys. In fact, it is not uncommon
to use a single mean scale factor for an entire project.
The grid factor, GF, for the line is a product of the elevation factor, EF,
and a scale factor, kavg, and is computed as
GF ⫽ kavg ⫻ EF
(F.60)
Thus, a reduced grid distance, Lgrid, is the product of the horizontal distance,
Lm, and the grid factor, GF, and is computed as
Lgrid ⫽ Lm ⫻ GF
(F.61)
Example F.1 A distance of 536.07 ft is observed from station 1. The scale
factors at observing, midpoint, and sighted stations are 0.9999587785,
0.9999587556, and 0.9999587328, respectively. The orthometric height at observing station is 1236.45 ft. Its geoidal height is ⫺30.12 m and the radius
in the azimuth is 6,366,977.077 m. Determine the length of the line on the
mapping surface.
592
MAP PROJECTION COORDINATE SYSTEMS
SOLUTION This solution will compare the grid factor computed using different radii in Equation (F.57) and different scale factors in Equation (F.60).
Using the more precise methods, the grid factor is computed as follows. The
orthometric height of the observing station in meters is
H ⫽ 1236.45 ft ⫻
12
⫽ 376.871 m
39.37
Using the radius in the azimuth of the line and Equation (F.57), the elevation
factor, EF, is
EF ⫽
6,366,977.077
⫽ 0.999945542
6,366,977.077 ⫹ 376.871 ⫺ 30.12
From Equation (F.59), the scale factor for the lines is
kavg ⫽
0.9999587785 ⫹ 4(0.9999587556) ⫹ 0.9999587328
⫽ 0.999958756
6
From Equation (F.60), the grid factor for the line is
GF ⫽ 0.999945542 ⫻ 0.999958756 ⫽ 0.99990430
Finally, the grid distance for this line is
Lgrid ⫽ 0.99990430 ⫻ 536.07 ft ⫽ 536.02 ft
Doing the problem again, this time with the mean radius of the Earth and the
average of the two endpoint scale factors, yields
EF ⫽
6,371,000
⫽ 0.999945577
6,371,000 ⫹ 376.871 ⫺ 30.12
kavg ⫽
0.9999587785 ⫹ 0.9999587328
⫽ 0.999958756
2
GF ⫽ 0.999945577 ⫻ 0.999958756 ⫽ 0.99990433
Lgrid ⫽ 0.99990433 ⫻ 536.07 ft ⫽ 536.02 ft
Note that using the approximate radius of the Earth and the average scale
factor for the endpoints of the line resulted in the same solution as the more
precise computations. This is because the length of the distance observed has
F.4
REDUCTION OF OBSERVATIONS
593
only five significant figures. The elevation factor computed using the mean
radius of the Earth agreed with the radius in the azimuth to seven decimal
places. This is also true of the scale factors, which agreed to nine significant
figures. Thus, the grid factor was the same to seven decimal places and was
well beyond the accuracy needed to convert a length with only five significant
figures. This demonstrates why a common grid factor can often be used for
an entire project that covers a small region.
F.4.2
Reduction of Geodetic Azimuths
Figure F.2 depicts the differences between geodetic azimuths, T, and grid
azimuths, t. Since grid north (GN) at a point is parallel to the central meridian,
the convergence angle, ␥, is the largest correction between the two geodetic
and grid azimuths. Additionally, there is a small correction to convert the arc
on an ellipsoid to its equivalent chord on the mapping surface. This is known
as the arc-to-chord correction, ␦. The relationship between the geodetic azimuth and grid azimuth can be derived from Figure F.2 as
T⫽t⫹␥⫺␦
(F.62)
As shown in Figure F.2, this equation works whether the line is east or west
of the central meridian. For the Lambert Conformal Conic map projection,
the arc-to-chord correction is computed as
␦ ⫽ 0.5(sin 3 ⫺ sin 0)(2 ⫺ 1)
(F.63)
An analysis of Equation (F.63) shows that the worst cases for ␦ are for lines
in the northern or southern extent of a map projection. Rearranging Equation
(F.63) yields a change in longitude as
GN
GN
Figure F.2 Relationship of geodetic azimuth (T), grid azimuth (t), convergence angle
(␥), and arc-to-chord correction (␦).
594
MAP PROJECTION COORDINATE SYSTEMS
2 ⫺ 1 ⫽ ⌬ ⫽
2
sin 3 ⫺ sin 0
(F.64)
As an example, assume that 3 is 42⬚30⬘. Further assume that the project is
in the Pennsylvania North Zone, which has a sin 0 of 0.661539733812. If ␦
is to be kept below 0.5ⴖ, the maximum line in arc-seconds of longitude can
be
⌬ ⫽
2(0.5ⴖ)
⫽ 71.2ⴖ
sin 42⬚30⬘ ⫺ 0.661539733812
At latitude 42⬚30⬘, this corresponds to a line of length of about 5334 ft, or
1.6 km. Few surveyors in northeastern Pennsylvania could find a line of this
length to observe. Thus, the arc-to-chord correction is generally ignored in
reductions, and Equation (F.62) can be simplified as
T⫽t⫹␥
(F.65)
APPENDIX G
COMPANION CD
G.1
INTRODUCTION
The companion CD that accompanies this book has several programs and
instructional worksheets to aid students in the learning process. The CD contains the software ADJUST, MATRIX, and STATS that were available with
the preceding edition of the book. Mathcad worksheets have been added to
this edition. For those who do not own Mathcad, html files of these worksheets have been created. These files can be viewed using your computer’s
Web browser.
All of the software contained on the CD is Windows-based and will run
on most Windows-based computers. The CD has an autoinstall program that
will allow you to select the packages for installation. Before installing the
software on the companion CD, it is important to remove any versions of the
software already on your computer. This can be done using the Windows
‘‘Add/Remove programs’’ feature in your Windows control panel. Refer to
the Windows help system if you are not familiar with this feature in Windows.
If the program fails to autoload on your machine, browse the CD and run the
‘‘setup.exe’’ file from the root directory of the CD. Each program contained
on the CD has different system memory requirements; although all will run
in less than 2 megabytes of memory.
This software is ‘‘freeware’’ and as such can be freely distributed with this
book. However, it is not intended for commercial use, is not guaranteed to
be computationally correct, and is not supported in any manner. It is simply
provided to aid your understanding of the topics contained in the book. As
the software is updated, it will be posted at http://surveying.wb.psu.edu under
the FREE GOODIES button. You should visit this site occasionally to download newer versions of the software on this CD.
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
595
596
G.2
COMPANION CD
FILE FORMATS AND MEMORY MATTERS
Most options in the software packages ADJUST, MATRIX, and STATS use
formatted ASCII text files for data entry. Each package is equipped with an
editor that allows the user to create these files. The help file for each software
program describes the format of the data files. It is also possible to create
data files using other ASCII editors, such as Notepad. The file-reading routines contained in ADJUST, MATRIX, and STATS can use either commas,
spaces, or tabs as delimiters between fields. It is therefore important to avoid
commas in large numbers such as coordinates, since the software will read
each part of the number individually. For example, a coordinate of
675,301.213 will be read as the numbers 675 and 301.213 since the comma
is a delimiter. The use of the TAB delimiter is especially useful in Matrix
since this allows you to ‘‘cut and paste’’ values from spreadsheets such as
Excel and Quattro Pro.
Starting with Version 4.0, ADJUST has used dynamic memory allocation.
That is, data storage structures are not created until their sizes have been
defined at runtime. Therefore, users familiar with earlier versions of ADJUST
should be aware that some file formats have changed to accommodate this
programming change. As a user, this feature means that the size limits of data
types in various options in the software are now limited only by your computer’s memory resources. Thus, much larger problems can be handled with
this software.
G.3
G.3.1
SOFTWARE
ADJUST
ADJUST is the main computational program on the companion CD. It contains programs that either support computation of problems in the book or
perform the computations. Several least squares programs and several supporting options are contained in ADJUST. A list of the least squares options
and supporting software is provided in Table G.1. For example, the traverse
option in ADJUST can compute initial approximations for unknown stations
in a horizontal plane survey. The ‘‘Horizontal Data’’ under ‘‘Least Squares
Adjustments’’ can perform the least squares adjustment of the data as discussed in Chapters 14 through 16.
File formats required by this software are discussed in the accompanying
Help file. At installation, sample data files of some of the numerical examples
presented in this book can be loaded onto your computer. By default, these
files are loaded in the ‘‘Adjustment Computations’’ subdirectory of the ‘‘MY
DOCUMENTS’’ directory. Along with the Help file, these sample data files
can be viewed and compared with the accompanying example problem in the
book to assist you in creating your own files.
G.3
SOFTWARE
597
TABLE G.1 Brief Summary of Software Options Contained in ADJUST
Option
Astronomical observations
Reduction for azimuth
Prediction of position
Coordinate computations
Forward
Inverse
Traverse
Area
State plane coordinates
Universal Transverse Mercator
Geodetic computations
Coordinate geometry
Geocentric coordinates
Oblique triangle solutions
Coordinate transformations
2D conformal
2D affine
2D projective
3D conformal and affine
Estimated errors
Horizontal / plane data
Differential leveling
Check errors
Fit of points
Line
Circle
Parabola
GPS data
Loop closure check
Baseline vector adjustment
Simulated adjustment
Least squares adjustments
Differential leveling data
Horizontal data
3D geodetic network
G.3.2
Data File Required
Least Squares
Adjustment
Yes
No
No
No
No
No
Yes
Yes
Dependent on option
Dependent on option
No
No
Dependent on option
No
No
No
No
No
No
No
No
No
No
No
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
No
No
No
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
No
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
STATS
STATS is a statistical package that computes the basic statistical properties
of simple data sets as well as deriving critical values for the normal, t, 2,
and F distributions.
598
COMPANION CD
G.3.3
MATRIX
MATRIX performs simple matrix operations. This software can be used to
solve many of the least squares problems in this book. It can read and write
files of matrices, perform the operations of scaling, transposition, addition,
subtraction, multiplication, and inversing of matrices, and allow the user to
view the results of these operations.
When solving least squares problems with this package, a spreadsheet can
be used to compute the matrices. These matrices can then be ‘‘cut and pasted’’
into the Matrix editor. Once saved to disk, the data files can be read by the
software and manipulated.
G.3.4
Mathcad Worksheets
The companion CD contains an electronic Mathcad book which will be installed in the HANDBOOK directory under Mathcad. The electronic book
can be opened from the Mathcad Help menu. For those who do not own
Mathcad, html files of the worksheets can be installed on your computer.
These files will be located in the ‘‘Mathcad HTML’’ directory in ‘‘MY DOCUMENTS.’’ The html files can be viewed using your HTML browser. After
installation, a link to these files may be found in the Programs menu.
The worksheets demonstrate most of the numerical examples contained in
this book. Most can be modified to compute other problems in the book.
Some of the worksheets read ASCII data files. These files can be created by
any ASCII editor, such as those contained in ADJUST, MATRIX, and STATS,
or by another package, such as Notepad. With many spreadsheets it is possible
to save a comma-separated values (.csv) text file. These files can also be read
by the worksheets. The format of the file is demonstrated by the numerical
example problem being solved in the original worksheet. Thus, when the
original worksheet is modified, it should be saved with a different name. You
should compare the file formats in these worksheets with the example problem
in the book to determine the proper format for your data.
These worksheets also provide a guide as to how various problems are
solved using programming. The Mathcad language is very similar to traditional programming languages. However, Mathcad does not provide global
variables or functions. Thus, all variables, especially those that will be modified, must be passed from function to function in a worksheet and between
worksheets. Mathcad reads the executable commands and variables from the
top to the bottom of the worksheet and from left to right on a line. This means
that location of a command or variable on the worksheet can be critical to
the worksheet performing properly. Another difference between a traditional
programming language and Mathcad is that Mathcad has both subscripted
variables and array elements. Unfortunately, although the subscript and array
element are entered differently, they look visually the same on the worksheet.
G.4
USING THE SOFTWARE AS AN INSTRUCTIONAL AID
Mathcad Azimuth Computation
Function:
Az(crds,i,1) :⫽ dx ← crdsj,1 ⫺ crdsi,1
dy ← crdsj,2 ⫺ crdsi,2
␣ ← atan2(dy,dx)
␣ ← ␣ ⫹ 2䡠 if ␣ ⬍ 0
return ␣
冨
599
C Language Azimuth Computation
Function:
double az(crdstruct crds[10], int i, int j)
{
double dx, dy, alpha;
dx ⫽ crds[j],[1] - crds[i],[1];
dy ⫽ crds[j],[2] - crds[i],[2];
alpha ⫽ atan2(dy,dx);
if (alpha⬍0) alpha ⫽ alpha ⫹ 2*PI();
return alpha;
}
Figure G.1 Comparison of a Mathcad function and a C function.
Thus, it is easy to confuse a subscripted variable with a matrix element. That
is, a0 may look like a reference to a matrix element when in fact it is a
subscripted variable. The user should refer to the Mathcad help system to
learn how to distinguish between, and use, subscripted variables and matrix
elements. The Mathcad code in Figure G.1 depicts references to matrix
elements.
Figure G.1 contains a function to compute the azimuth of a line based on
the coordinates of the endpoints. Except for the slight language-specific differences, the code in the Mathcad and C functions is very similar. Thus, with
some modifications, the code in the Mathcad worksheets can serve as a model
when developing similar code in a traditional programming language.
G.4
USING THE SOFTWARE AS AN INSTRUCTIONAL AID
Many of the problems presented in this book can be solved using the software
on the companion CD. However, it would not be wise to solve all problems
with this software, since true understanding can only be gained by solving
the problem yourself. Still, some problems are so repetitive or long that it is
extremely difficult to solve them correctly without the aid of software. In
these cases, the reader is often referred to the software on the companion CD.
For example, in Appendixes A through C, the matrix operations of addition,
subtraction, multiplication, and inversing are presented. Although the student
is expected to solve the accompanying problems in these appendices by hand,
the MATRIX program should be used to solve the remainder of the problems
presented in this book. This frees the reader of the matrix operations so that
emphasis can be placed on the topics presented in the chapters. However, this
software can be used to check solutions. The Mathcad worksheets are extremely valuable for this since the intermediate steps in the solution can also
be viewed and checked against written work.
BIBLIOGRAPHY
Amer, F. 1979. Theoretical Reliability Studies for Some Elementary Photogrammetric
Procedures, Aerial Triangulation Symposium, Department of Surveying, University
of Queensland, St. Lucia, Australia.
Baarda, W. 1967. Statistical Concepts in Geodesy, Netherlands Geodetic Commission,
Delft.
———. 1968. A Testing Procedure for Use in Geodetic Networks, Netherlands Geodetic Commission.
Bjerhammar, A. 1973. Theory of Errors and Generalized Matrix Inverses, Elsevier
Science, New York.
Box, George E. P., et al. 1978. Statistics for Experimenters, Wiley, New York.
Bomford, G. 1980. Geodesy, 4th ed., Claredon Press, Oxford.
Bruns, H. 1878. Die Figur der Erde, Preuss. Geod. Inst., Berlin.
Buckner, R. B. 1983. Surveying Measurements and Their Analysis, Landmark Enterprises, Rancho Cordova, CA.
Burse, Michelle L. 1995. Profile of a Least Squares Convert, Point of Beginning, 20(2):
76–82.
Conte, S. D., and Carl de Boor. 1980. Elementary Numerical Analysis, 3rd ed.,
McGraw-Hill, New York.
Dewitt, Bon A. 1994. An Efficient Memory Paging Scheme for Least Squares Adjustment of Horizontal Surveys, Surveying and Land Information Science, 54(3):
147–156.
Dracup, Joseph F. 1994. Squares Adjustment by the Method of Observation Equations
with Accuracy Estimates, Surveying and Land Information Science, 55:2.
El-Hakim, S. F. 1981. A Practical Study of Gross-Error Detection in a Bundle Adjustment, Canadian Surveyor, 35(4):373–386.
———. 1984. On the Detection of Gross and Systematic Errors in Combined Adjustment of Terrestrial and Photogrammetric Data, Commission III, International Archives of Photogrammetry and Remote Sensing, pp. 151–163.
———. 1986. The Detection of Gross and Systematic Errors in Combined Adjustment
of Terrestrial and Photogrametric Data, Photogrammetric Engineering and Remote
Sensing, 52(1):59–66.
600
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
BIBLIOGRAPHY
601
Fubara, D. M. J. 1972. Three-Dimensional Adjustment of Terrestrial Geodetic Networks, Canadian Surveyor, 26:4.
George, Alan, and Joseph W.-H. Liu. 1981. Computer Solution of Large Sparse Positive Definite Systems, Prentice-Hall, Englewood Cliffs, NJ.
Ghilani, Charles D. 1990. A Surveyor’s Guide to Practical Least Squares Adjustments,
Surveying and Land Information Science, 50(4):287–297.
———. 1994. Some Thoughts on Boundary Survey Measurement Standards, Surveying
and Land Information Science, 54(3):161–167.
———. 2003a. Statistics and Adjustments Explained, Part 1: Basic Concepts, Surveying and Land Information Science, 63(2):73.
———. 2003b. Statistics and Adjustments Explained, Part 2: Sample Sets and Reliability,’’ Surveying and Land Information Science, 63(3):141.
———. 2004. Statistics and Adjustments Explained, Part 3: Error Propagation, Surveying and Land Information Science, 64(1):29–33.
Hartzell, P., L. Strunk, and C. Ghilani. 2002. Pennsylvania State Plane Coordinate
System: Converting to a Single Zone, Surveying and Land Information Science,
62(2):95–103.
Harvey, Bruce R. 1994. Practical Least Squares and Statistics for Surveyors, School
of Surveying, University of New South Wales, Australia.
Heiskanen, W. A., and H. Moritz. 1967. Physical Geodesy. W. H. Freeman and Company, San Francisco.
Hirvonen, R. A. 1965. Adjustment by Least Squares in Geodesy and Photogrammetry,
Frederick Ungar, New York.
Hoffman-Wellenhof, B., et al. 2001. GPS Theory and Practice, 5th ed., SpringerVerlag, New York.
Hogg, Robert V., and Johannes Ledolter. 1992. Applied Statistics for Engineers and
Physical Scientists, Macmillan, New York.
Hotine, M. 1959. A Primer of Non-classical Geodesy, Presented to the First Symposium
on Three-Dimensional Geodesy, Venice.
Kuang, Shanlong. 1994. A Strategy for GPS Survey Planning: Choice of Optimum
Baselines, Surveying and Land Information Science, 54(4):187–201.
Leick, Alfred. 2004. GPS Satellite Surveying, 3rd edition, Wiley-Interscience, New
York.
McMillan, Kent Neal. 1995a. Least Squares: Older and Better Than Barbed Wire,
Point of Beginning, 20(2):82–84.
———. 1995b. Least Squares Under the Hood, Point of Beginning, 20(2):84–88.
Mendenhall, William, and Terry Sincich. 1994. Statistics for Engineering and the Sciences, Dellen Publishing, San Francisco.
Mikhail, Edward M. 1976. Observations and Least Squares, University Press of America, Washington, DC.
Mikhail, Edward M., and Gordon, Gracie. 1981. Analysis and Adjustment of Survey
Measurements, Van Nostrand Reinhold, New York.
Millbert, Kathryn O., and Dennis G. Milbert. 1994. State Readjustments at the National
Geodetic Survey, Surveying and Land Information Science, 54(4):219–230.
Misra, Pratap, and Per Enge. 2001. Global Positioning System: Signals, Measurements,
and Performance, Ganga-Jamuna Press, MA.
602
BIBLIOGRAPHY
Schwarz, Charles R. 1994. The Trouble with Constrained Adjustments, Surveying and
Land Information Science, 54(4):202–209.
Schwarz, K. P., E. H. Knickmeyer, and H. Martell. 1990. Assessment of Observations
Using Minimum Norm Quadratic Unbiased Estimation, CISM Journal ACSGC,
44(1):29–37.
Seber, G. A. F. 1977. Linear Regression Analysis, Wiley, New York.
Sideris, Michael G. 1990. The Role of the Geoid in One, Two-, and Three-Dimensional
Network Adjustments, CISM Journal ACSGC, 44:1:9–18.
Snay, Richard A. 1976. Reducing the Profile of Large Sparse Matrices, NOAA Technical Memorandum NOS-NGS 4, republished in Bulletin Geodésiqué, 50(4):341.
Strang, G., and Kai Borre. 1997. Linear Algebra, Geodesy, and GPS, Wellesley–
Cambridge Press, Wellesley, MA.
Vanicek, P., and E. Krakiwsky. 1992. Geodesy: The Concepts,. Elsevier, New York.
Veis, G. 1960. Geodetic Use of Artificial Satellites, Smithsonian Contributions to Astrophysics. 3(9):95–161.
Vincenty, T. 1979. The HAVAGO Three-Dimensional Adjustment Program. NOAA
Technical Memorandum NOS NGS 17. National Technical Information Service,
NOAA, Silver Spring, MD.
White, L. A. 1987. Calculus of Observations 380 / 580, Western Australian Institute of
Technology, Perth, Australia.
Wolf, Paul R., and Charles D. Ghilani. 2006. Elementary Surveying, 11th ed., Prentice
Hall, Upper Saddle River, NJ.
INDEX
Accuracy, 4, 5
example of, 5, 6
Adjoints, 537
ADJUST, 596
options, 597
Adjustment
analysis of, 492
conditional, 178, 196
constrained, 64, 78, 388, 395, 398,
400, 412
of control stations, 388, 395, 412,
418, 468
free, 291, 326
fully constrained, 64, 291, 412,
419
geodetic network, 454
leveling, 205
minimally constrained, 64, 291,
326, 412, 418
networks, 291–300
parametric, 178
traverse, 283
triangulation, 255
trilateration, 233
Alternative hypothesis, 68
Analysis of adjustments, 492
Arc-to-chord correction, 593
Associative law, 526
Astronomical observations for
azimuth, 116
Atmospheric refraction, 472
Azimuth
astronomical observations for, 116
constraint, 295
equation for, 255, 603
errors in astronomical observations
for, 116
Baarda, 416
Back-substitution, 511
Blunder detection
a posteriori, 406, 412, 416, 420,
492–502
a priori, 322–326, 410–412
data snooping, 416, 494
example of, 420–428, 493–496
graphical methods, 411, 499
handling of control, 326, 388, 418,
468
network, 291, 327, 409, 492
numerical example, 420, 493
procedures, 418, 492
residuals, 413, 492
by residual sign, 413, 492, 494
statistical methods, 44, 416, 501
Adjustment Computations: Spatial Data Analysis, Fourth Edition. C. D. Ghilani and P. R. Wolf
© 2006 John Wiley & Sons, Inc. ISBN: 978-0-471-69728-2
603
604
INDEX
Blunder detection (continued )
traverse closure checks, 114, 130,
135, 411
by traverse closures, 114, 127, 130,
411
Bomford, 474
Bruns, 454
Calibration, EDM, 32, 73, 195
Centering error
instrument, 100, 106, 121
target, 100, 104, 121
Central meridian, 583, 587
2 distribution, 52, 564
goodness of fit test, 300, 412, 496
test, 74, 300, 412, 496
use of table, 62, 74
Cholesky decomposition, 508
Class
example of use, 27
frequency, 15
interval, 15, 16
width, 14
Closure check, angular, 114
Cofactor, definition of, 537
Companion CD, 595
ADJUST, 596
file formats, 596
MATRIX, 598
STATS, 597
Computer optimization, 504
processing, 507, 508, 513
storage, 504, 513
Confidence interval, 50
interpretation of, 59
for mean, 56, 59
for population variance, 61
ratio of two population variances,
63
variance, 61
Confidence interval for mean,
computation of, 58
Confidence interval for population
variance, computation of, 62
Constraint equations, 388
azimuth example, 395
control coordinates, 388, 395, 468
differential leveling example, 398
by elimination, 394
geometric, 173
Helmert’s method, 398
Control, adjustment of, 139, 326,
388, 418, 468
Conventional Terrestrial Pole, 316,
480
Coordinate systems
geocentric, 316, 481
geodetic, 316, 478
global, 478, 480
local, 478, 480
local geodetic, 455
NAD 83, 316, 478
nearly aligned, 479
rotations between, 480
satellite, 314
state plane coordinate system, 233,
582
WGS 84, 316, 318, 478, 479
Coordinate transformations
adjustment of, 345
datums, 480
eight parameter, 353, 448
four parameter, 345, 444
Helmert, 480
rotations, 357
scaling, 346, 351, 356
seven parameter, 356, 449
six parameter, 350, 447
statistically valid parameters, 362
three-dimensional conformal, 356,
449
two-dimensional conformal, 345,
444
two-dimensional affine, 350, 447
two-dimensional projective, 353,
448
Coordinates
geocentric, 316, 481
geodetic, 316, 478
global, 478, 480
local, 478, 480
local geodetic, 455
state plane, 233, 582
Covariance
definition of, 86
development of matrix, 221
matrix, 159
INDEX
Data
classes, 14
frequency of, 14
graphical representation, 14
measures of central tendency, 17
measures of relative standing, 17
measures of variation, 17, 19
numerical methods of describing,
17
range of, 13
skewed, 17
Datum
earth-centered, earth-fixed, 480
global, 478, 480
International Terrestrial Reference
Frame, 478
local, 478, 480
North American Datum of 1983,
316, 478
transformation of, 480
WGS84, 316, 318, 478, 479
Deflection of the vertical, 472
Degrees of freedom
in an adjustment, 212
constrained adjustment, 403
definition of, 19
in a sample set, 53
Density function, normal, 36
Determinant, 535, 538
Differential leveling, weights in, 166
DIN 18723, angular error, 103
Discrepancy, 4
Dispersion, 13
Distribution
bivariate, 109, 369
2, 52, 564
F, 55, 568
t, 54, 566
Distribution function
normal, 35, 36
standard normal, 38, 39
Eccentricity, 316, 317, 583, 587
EDM, error sources, 121
Elevation factor, 590
Elimination of constraints, 394
numerical example, 395
605
Ellipse
example computation of, 376, 377
eccentricity, 317, 583, 587
flattening factor, 318, 479
orientation of, 371, 375
quu, 375
qvv, 376
semimajor axis, 376
semiminor axis, 376
t angle, 374
Ellipsoid, 478
definition of, 479
eccentricity, 317, 583, 587
flattening factor, 318, 479
GRS 80, 318, 478, 479, 583
radius in the meridian, 461, 468,
484
radius in normal, 318, 461, 468,
484
WGS 84, 316, 318, 478, 479
Equations
conditional, 178, 196
normal, 181, 184, 185
observation, 179
Error
blunder, 4, 409, 417
definition of, 3, 18
ellipse, 369, 375–378
gross error, 4
instrument, 3
natural, 3
percent probable, 43
personal, 3
random, 4
standard, 20, 38
systematic, 4
Type I, 69, 416
Type II, 69, 416
Error ellipse, 369, 427
advantages, 381
confidence level, 379
drawing of, 378
numerical example of, 376, 377
percent probability, 379
Error propagation, 2, 84, 99
in adjusted quantities, 225
angle, 100–112
astronomical observation, 116
azimuths, 169
606
INDEX
Error propagation (continued )
in computed quantities, 226
covariance matrix, 87, 221
differential leveling, 148
EDM, 121
intersection example, 91
numerical examples, 89
standard error in a series, 89
standard error of a sum, 88
standard error of mean, 89
traverse, 127
trigonometric leveling, 152
Error sources
instrumental, 3
natural, 3
personal, 3
random, 4
systematic, 4
Estimators, definition of, 52
False easting, 583, 587
False northing, 583, 587
F distribution, 568
definition of, 55
Fisher distribution, definition of, 52
Fit of points
circle, 550–554
line, 192, 437
parabola, 194
Forward substitution, 508
Free network adjustment, 291, 326
Frequency
class, 15
histogram, 15
Fundamental principle of least
squares, 174
Gauss, 173
General law of propagation of
variances, definition of, 87
General least squares method, 437
Geocentric coordinates, 316
Geocentric coordinate system, 316,
481
Geodetic network
adjustment of, 454, 471
azimuth observations, 457
differential leveling, 460
horizontal angle, 459
horizontal distances, 460
linearization of equations, 456
minimum number of constraints,
462
slant distances, 457
vertical angle, 459
Geodetic Reference System of 1980,
316, 478
Geographic information systems, 1
Geoid model, 462
Geoidal height, 473
GLOPOV, definition of, 87
Goodness-of-fit test, 300, 412, 496
GPS, 310, 478
A matrix, 327
baselines, 312
carrier phase measurements, 312
differencing, 312
errors, 314
least squares adjustment, 310, 478
observations, 311
preadjustment data analysis, 322
real-time kinematic survey, 501
reference coordinate systems, 314
weight matrix, 329
Graphical representation, histogram,
14
Grid origin, 583, 587
Heights
geodetic, 316, 473
geoid, 316, 462
geoidal height, 316, 462, 473
orthometric, 316, 473
Histogram, 14
bimodal, 17
classes, 14
class width, 14
common shapes, 17
definition of, 14
example construction, 23
skewed to the right, 17
Horizontal angle
error sources, 99
instrumental centering error, 106
leveling error, 110
INDEX
linearized observation equation,
259, 459
pointing error, 102
reading error, 100
target centering error, 104
Horizontal time-dependent
positioning, 483
Hypothesis testing
alternative hypothesis, 68
definition of, 68
null hypothesis, 68
population mean, 72
population variance, 74
ratio of two population variances,
77
rejection region, 69
test statistic, 68
Hypothesis
alternative, 68
null, 68
testing, 68
Inaccessible point, example of, 91
Instructional aids, 595, 599
ADJUST, 596
Mathcad worksheets, 598–599
MATRIX, 598
STATS, 597
Instrument centering error, example
of, 109
Intersection
angle, 260
trilateration, 237
Inverse
by adjoints, 537
by row transformations, 538
of 2 by 2 matrix, 535
Iteration termination, 248
by change in reference variance,
249
by correction size, 249
by maximum iteration, 249
Jacobian matrix, 188, 549
Lambert conformal conic, 583
direct problem, 585
607
inverse problem, 585–586
zone constants, 584
Laplace, 173
Least squares adjustment
control coordinates, 388
coordinate transformation, 345, 444
differential leveling, 205
general, 437
geodetic network, 454
GPS, 310, 478
line fit of points, 191
nonlinear system of equations, 188
parabola, 194
traverse, 283
triangulation, 255
trilateration, 233
Legendre, 173
Leveling
adjustment of, 205
in angle measurement, 110
collimation error in, 144
Earth curvature and refraction, 146
random errors in, 148
in rod plumbing, 148
systematic errors, 144
Linearized observation equation
angle, 259, 465
azimuth, 258, 467
distance, 237
slant distance, 464
vertical angle, 466
Link traverse
estimated closure of, 136
Local geodetic system, 454
Localization, 483
Map projection, 582–583
elevation factor, 590
grid factor, 590
Lambert conformal conic, 583
reduction of azimuths, 593
reduction of distances, 590
reduction of observations, 590
scale factor, 590–591
Universal Transverse Mercator,
586–587
Mapping table, 506
Mathcad worksheets, 598–599
608
INDEX
Mathematical model
conditional, 178
parametric, 178
Matrices, 520
addition of, 524
column, 522
covariance, 87, 221
definition of, 520
diagonal, 523
dimensions of, 521
equality of, 523
idempotent, 415
identity, 523
inverse of, 534
Jacobian, 188, 549
lower triangular, 508
multiplication by a scalar, 524
multiplication of, 525
partitioning, 229
rectangular, 522
rotation, 357
row, 522
spareness and optimization, 513
square, 522
subtraction of, 524
symmetric, 523
systematic formation, 185, 507
transpose, 523
unit, 523
upper-triangular, 508
Mean
confidence interval of, 56
definition of, 18
example computation, 23, 25
standard deviation in, 21
weighted, 160, 161
Measurements
direct, 2
indirect, 2
Measures of central tendency, 17
definition of, 17
mean, 18
median, 13, 18
Median
definition of, 18
example computation, 23
Meridian, radius in the, 461, 468,
484
Minimally constrained adjustment,
64, 291, 326, 412, 418
Mistake, definition of, 3
Mode
definition of, 18
example computation, 23
Model
stochastic, 177
weighting, 159
Most probable value, 19, 174
National Geodetic Survey, 472, 474,
582
National Geospatial-Information
Agency, 586
Network
adjustment of, 291
design of, 381, 430
example of adjustment, 292
Nonlinear equations, solution of, 188,
546, 549
Normal, radius of, 318, 461, 468,
484
Normal distribution
function, 38
general principles, 35, 492
inflection point, 38, 41
Normal equations,
formation of, 181, 185
matrix formation, 185
solution, 187, 189
tabular formation, 184
Normal error distribution curve, 35
North American Datum
1927, 345
1983, 234, 345, 478–480
Null hypothesis, 68
Observation equation
angle, 258, 459
azimuth, 255, 457
control coordinates, 388
definition of, 179
differential leveling, 205, 460
distance, 235, 460
slant distance, 457
vertical angles, 459
INDEX
Observation weight, 160
Optimization
direct formation of normal
equations, 507
storage, 513
Orthometric height, 316, 473
Outliers, detection of, 42, 44, 59, 416
Percent errors, uses of, 43
Percentage points, definition of, 53
Pointing and reading errors
DIN 18723, 103
example of, 101–103, 112
Population
definition of, 13
sampling, 12
standard deviation, 20
variance, 19, 74
Precision
definition of, 4
example of, 5
Probability
definition of, 33
in compound event, 33
joint, 34
Probable error
example of, 44
50%, 42
95%, 42
other percent errors, 43
standard error, 41
table of, 43
use of, 45, 46
Propagation, errors, 84, 221
Pseudorandom noise, 311
Pseudorange, 311, 499
Radius in the meridian, 461, 468,
484
Random error, definition of, 4
Random samples, 50
Range, 13, 16
example of, 23, 25
Ratio of variances, confidence
interval of, 63, 74
Reading error
horizontal angle, 101–103, 112
609
leveling, 148
Redundancies, see Degrees of
freedom
Redundancy number, 416, 428
Reference standard deviation
computation of, 212, 213
definition of, 211
Reference standard deviation of unit
weight, 161
Reference variance
definition of, 160
goodness-of-fit test, 300, 412, 496
Rejection region, 69
Reliability
external, 429
internal, 429
Resection
initial approximations, 265
least squares adjustment of, 270
Residual
analysis of, 442
comparison of, 493, 500
definition of, 19
plots, 500
RTK-GPS, 501
Sample
definition of, 12
distribution theory, 52
mean, 18, 23, 25
size, 41, 60
variance, 19, 20, 21, 52
Satellite orbit
apogee, 314
argument of perigee, 316
ascending node, 316
line of apsides, 314
perigee, 314
right ascension, 316
Site calibration, 483
Slant distance, 455, 457
SLOPOV
definition of, 88
example of, 88, 127
Software
ADJUST, 596
Mathcad worksheets, 598–599
610
INDEX
Software (continued )
MATRIX, 598
STATS, 597
spreadsheets, 123
Stakeout surveys, RTK-GPS, 501
Standard deviation
for computed quantities, 226
definition of, 20
example computation, 23, 27
mean, 21, 89
sample, 20
series, 89
sum, 88
Standard error
definition of, 20
probability of, 41
Standard parallel, 583
State plane coordinate system, 233,
582
Statistics, 17, 52
STATS, 597
Stochastic model, 177
importance of, 499
Student distribution, definition of, 54,
562
Survey design, 381, 430
field specifications, 431
procedures, 430
use of maps, 430
uses of, 94
Systematic errors
in angles, 472
collimation, 144
definition of, 4
in heights, 473
in leveling, 144
Tables
2 distribution, 565
F distribution, 569–575
standard normal distribution, 562–
563
t distribution, 567
Target centering error
example of, 101, 104, 121
Taylor series approximation, 546
t distribution, 54, 566
definition of, 54
use of, 56, 72, 114, 131, 134, 363,
416, 493
Terrestrial observation
with GPS baseline vectors, 478
Test
2, 74, 300, 412, 496
F, 55, 419
goodness of fit, 300, 412, 496
hypothesis, 68
population mean, 19, 72
population variance, 74
ratio of population variances, 77
significance of, 70
statistic, 68
t, 72
Three-dimensional coordinate
transformation
conformal, 356
Total station, DIN 18723, 103
Transformation
Helmert, 480
three-dimensional conformal, 356,
449
two-dimensional affine, 350, 447
two-dimensional conformal, 345,
444
two-dimensional projective, 353,
448
TRANSIT, 310
Transverse Mercator, 586–587
direct problem, 588
inverse problem, 588–589
zone constants, 587
Traverse
estimated closure, 128, 135
minimally constrained, 291
minimum control, 291
numerical example, 128, 136, 285
redundant equations, 135, 284
Traverse adjustment
degrees of freedom, 284
example of, 285
observation equations, 284
redundancies, 284
Traverse closure
anticipated, 128, 136
numerical example, 128, 136
Triangular matrices, forward and
backward solution, 511
INDEX
Triangulation adjustment, 255
Trilateration adjustment, 233
True value, 3, 18
Two-dimensional coordinate
transformation
affine, 350, 447
conformal, 356, 443
projective, 353, 448
Type I error, 69, 416
Type II error, 69, 416
Universal Transverse Mercator, 586–
587
Variance
confidence interval of, 61
611
definition of, 19, 22
population, 19
sample, 19
Vincenty, 456
Weighted mean, 161
angles, 165
example of, 168
leveling, 166
Weight, 159
constrained adjustment, 403
relationship to corrections, 159
relationship to variances, 160
uncorrelated observations, 160
uses of, 159
World Geodetic System of 1984, 316,
318, 478, 479