As discussed in Section 1.2, unknown values are often determined indirectly by making direct measurements of other quantities that are functionally related to the desired unknown quantities. Examples in surveying include computing station coordinates from distance and angle observations, obtaining station elevations from rod readings in differential leveling, and determining the azimuth of a line from astronomical observations. As noted in Section 1.2, since all directly observed quantities contain errors, any values computed from them will also contain errors. This intrusion, or propagation, of errors that occurs in quantities computed from direct measurements is called error propagation. This topic is one of the most important discussed in this book.
In this chapter, it is assumed that all systematic errors and mistakes have been eliminated from a set of direct observations, so that only random errors remain. To derive the basic error propagation equation, consider the simple function, z = a1x1 + a2x2, where x1 and x2 are two independently observed quantities with standard errors σ1 and σ2, and a1 and a2 are constants. By analyzing how errors propagate in this function, a general expression can be developed for the propagation of random errors through any function.
Since x1 and x2 are two independently observed quantities, they each have different probability density functions. Let the errors in n determinations of x1 be , and the errors in n determinations of x2 be , then zT, the true value of z for each independent observation, is
The values for z computed from the observations are
Substituting Equations (6.2) into Equations (6.1) and regrouping Equations (6.1) to isolate the errors for each computed value yields
From Equation (2.4) for the variance in a population, , and thus for the case under consideration, the sum of the squared errors for the value computed is
Expanding the terms in Equation (6.4) yields
Factoring terms in Equation (6.5) results in
Inserting summation symbols for the error terms in Equation (6.6) yields
Recognizing that the terms in parentheses in Equation (6.7) are by definition: , respectively, Equation (6.7) can be rewritten as
In Equation (6.8) the middle term, , is known as the covariance. This term shows the interdependence between the two unknown variables x1 and x2. As the covariance term decreases, the interdependence of the variables also decreases. When these terms are zero, the variables are said to be mathematical independent. Its importance in surveying is discussed in more detail in later chapters.
Equations (6.7) and (6.8) can be written in matrix form as
where Σzz is the variance-covariance matrix for the function z. It follows logically from this derivation that, in general, if z is a function of n independently measured quantities, x1, x2,…, xn, then Σzz is
Further for a set of m functions with n independently measured quantities, x1, x2,…, xn, Equation (6.10) expands to
Similarly, if the functions are nonlinear, a first order Taylor Series expansion can be used to linearize them.1 Thus, a11, a12,…are replaced by the partial derivatives of the Z1, Z2,…with respect to the unknown parameters, x1, x2,…. Thus, after linearizing a set of nonlinear equations, the matrix for the function of Z can be written in linear form as
Equations (6.11) and (6.12) are known as the general law of propagation of variances (GLOPOV) for linear and nonlinear equations, respectively. Both Equations (6.11) and (6.12) can be written symbolically in matrix notation as
where Σzz is the covariance matrix for the function Z. For a nonlinear set of equations that is linearized using Taylor's Theorem, the coefficient matrix (A) is called a Jacobian matrix. That is, it is a matrix of partial derivatives with respect to the unknowns, as shown in Equation (6.12).
If the observations, x1, x2,…, xn, are unrelated, that is, they are statistically independent, then the covariance terms are equal to zero and the right side of Equations (6.10) and (6.11) can be rewritten, respectively, as
If there is only one function Z, involving n unrelated quantities, x1, x2,…, xn, then Equation (6.15) can be rewritten in algebraic form as
Equations (6.14) (6.15), and (6.16) express the special law of propagation of variances (SLOPOV). These equations govern the manner in which errors from statistically independent observations (i.e., ) propagate in a function. In these equations, individual terms represent the individual contributions to the total error that occur as the result of observational errors for each independent variable. When the size of a function's estimated error is too large, inspection of these individual terms will indicate the largest contributors to the error. The most efficient method to reduce the overall error in the function is to closely examine ways to reduce the largest individual error terms in Equation (6.16).
Let A = B + C, and assume that B and C are independently observed quantities. Note that and . Substituting these into Equation (6.16) yields
Using Equation (6.15) yields
Equation (6.17) yields the same results as Equation (6.16) after the square root of the single element is determined. In the equations above, standard error (σ) and standard deviation (S) can be used interchangeably.
Let A = B1 + B2 + ⋯ + Bn, where the B's are n independently observed quantities having standard deviations of , then by Equation (6.16),
Assume that the error for each observed value in Equation (6.18) is equal—that is, , then Equation (6.18) simplifies to
Let be the mean obtained from n independently observed quantities y1, y2,…, yn, each of which has the same standard deviation S. As given in Equation (2.1), the mean is expressed as
An equation for , the standard deviation of the mean is obtained by substituting the above expression into Equation (6.16). Since the partial derivatives of with respect to the observed quantities, y1, y2,…, yn, is , the resulting error in is
Note that Equation (6.20) is the same as Equation (2.8).
The computations in this chapter can be time-consuming and tedious often leading to computational errors in the results. It is often more efficient to program these equations in a computational package. The programming of the examples in this chapter is demonstrated in the Mathcad® electronic book on the companion website for this book. Figure 6.4 shows a partial listing of a Mathcad® worksheet used to solve Example 6.3. Notice that a function called radian was used to convert angles from sexagesimal units to radian units. After all of the variables for the problem are created, the problem is solved by creating an equation that looks very similar to those written in this book. The results of each expression are shown to the right of the equation. Oftentimes, the time that is spent learning a new programming language is more than rewarded in its efficient use to solve multiple problems. For example, the same worksheet can be used to solve problems that are similar to that in Example 6.3.
Another efficient method of programming these types of problems is to use a spreadsheet. In some spreadsheets, cells can be named, and thus equations can be entered using variables. For example, the expression for the , which is EC/A in Figure 6.5, was entered as AB/2*((−COS(A+B)*(SIN(B)*TAN(VA)+SIN(A)*TAN(VB)))/(SIN(A+B))^2 + (COS(A)*TAN(VB))/SIN(A+B)) where AB referred to the cell containing the distance AB, A the cell containing the angle A in radian units, B the cell containing the angle B in radian units, VA the cell containing the altitude angle v1 in radian units, and VB the cell containing the altitude angle v2 in radian units. Similarly, the other expressions were created using the named cells. While this method is not as clear as using a program such as Mathcad®, when an expression is entered carefully, the same results can be computed. Again, the resultant spreadsheet can be used to solve similar problems by simply change the data entry values.
Errors associated with any indirect measurement problem can be analyzed as described above. Besides being able to compute the estimated error in a function, the sizes of the individual errors contributing to the functional error can also be analyzed. This identifies those observations whose errors are most critical in reducing the functional error. An alternative use of the error propagation equation involves computing the error in a function of observed values prior to fieldwork. The calculation can be based on the geometry of the problem and the observations that are included in the function. The estimated errors in each observed value can be varied to correspond with those estimated using different combinations of available equipment and field procedures. The particular combination that produces the desired accuracy in the final computed function can then be adopted in the field. This analysis falls under the heading of survey planning and design. This topic is discussed further in Chapters 7, 19, and 21.
Note: Partial answers to problems marked with an asterisk can be found in Appendix H.
Section | Observed Length (m) | Standard Deviation (mm) |
AB | 323.532 | ±3.2 |
BC | 465.083 | ±3.3 |
CD | 398.706 | ±3.2 |
DE | 120.683 | ±3.0 |
AB = 203.56 ± 0.02 ft | |
What is the elevation of the stack and the standard deviation in this elevation?
ElevA = 248.35 ± 0.04 ft | ElevB = 242.87 ± 0.04 ft |
hiA = 5.53 ± 0.01 ft | hiB = 5.52 ± 0.01 ft |