Acessibilidade / Reportar erro

Uncertainties Associated with Arithmetic Map Operations in GIS

Abstract

Arithmetic map operations are very common procedures used in GIS to combine raster maps resulting in a new and improved raster map. It is essential that this new map be accompanied by an assessment of uncertainty. This paper shows how we can calculate the uncertainty of the resulting map after performing some arithmetic operation. Actually, the propagation of uncertainty depends on a reliable measurement of the local accuracy and local covariance, as well. In this sense, the use of the interpolation variance is proposed because it takes into account both data configuration and data values. Taylor series expansion is used to derive the mean and variance of the function defined by an arithmetic operation. We show exact results for means and variances for arithmetic operations involving addition, subtraction and multiplication and that it is possible to get approximate mean and variance for the quotient of raster maps.

Key words
GIS; Interpolation variance; Map Algebra; Propagation of Uncertainty; Taylor series

INTRODUCTION

A map resulting from interpolation of field data must have some assessment of uncertainty (e.g. Heuvelink et al. 1989, Crosetto et al. 2000, Atkinson and Foody 20023 CROSETTO M, TARANTOLA S AND SALTELI A. 2000. Sensitivity and uncertainty analysis in spatial modelling based on GIS. Agr Ecosyst Environ 81: 71-79.). When estimates are derived from field data, they have associated uncertainties caused by spatial variation of continuous variables, interactions among them and the effect of neighbor data (Wang et al. 200513 WANG G, GERTNER GZ, FANG S AND ANDERSON AB. 2005. A methodology for spatial uncertainty analysis of remote sensing and GIS products Photogramm Eng Rem S 71: 1423-1432.). Generally, kriging is used to predict values of the variable of interest at unsampled locations, because it provides an assessment of uncertainty as given by the kriging variance. However, the kriging variance does not depend on real local data values (Atkinson and Foody 20021 ATKINSON PM AND FOODY GM. 2002. Uncertainty in remote sensing and GIS: fundamentals. In: Foody GM and Atkinson PM (Eds), Uncertainty in Remote Sensing and GIS, p. 1-18.) and therefore this is not a reliable measure of local accuracy. According to Wang et al. (2005), local estimates are strongly associated with neighbor data. Actually, the kriging variance is just a measure of the ranking of data configurations given by the variogram model (Journel and Rossi 19898 JOURNEL AG AND ROSSI M. 1989. When do we need a trend model in kriging? Math Geolo 21: 715-739.). A measure of local accuracy should take into account the data value. In this sense, Yamamoto (200016 YAMAMOTO JK. 2000. An alternative measure of the reliability of ordinary kriging estimates. Math Geol 32: 489-509.) proposed interpolation variance as a measure of the reliability of kriging estimates taking into account both data configuration and data values.

Moreover, interpolation variance can also be used for quantifying uncertainty associated with interpolation of categorical variable (Yamamoto et al. 201218 YAMAMOTO JK, MAO XM, KOIKE K, CROSTA AP, LANDIM PMB, HU HZ, WANG CY AND YAO LQ. 2012. Mapping an uncertainty zone between interpolated types of a categorical variable. Comput Geosci 40: 146-152.). Therefore, interpolation variance is valid for both continuous and discrete variables. Actually, the interpolation variance can be used with any interpolation method based on weighted average formula (for instance, inverse of distance weighting, multiquadric equations, etc.). Yamamoto et al. (2012) used interpolation variance as a measure of uncertainty associated with multiquadric interpolation. Because interpolation is based on limited information given by a sample, the resulting raster map is always subject to uncertainty. Usually, the main source of uncertainty is related to a lack of knowledge, but there are other factors affecting the final uncertainty in a raster map. Burrough (1986) lists 14 factors affecting the uncertainty. These factors were classified into three groups: I) obvious sources of error; II) errors resulting from natural variation of original measurements; III) errors arising through processing (Burrough 19862 BURROUGH PA. 1986. Principles of geographical information system. Oxford, Oxford University Press, 194 p.). Wellmann et al. (201015 WELLMANN FJ, HOROWITZ GF, SCHILL E AND REGENAUER-LIEB K. 2010. Towards incorporating uncertainty of structural data in 3D geological inversion. Tectonophysics 490: 141-152.) classified sources of uncertainty affecting geological data into three categories: I) imprecision and measurement error; II) stochastic nature of the geological variable; III) imprecise knowledge. Considering the field data with negligible uncertainty, the measured variance at unsampled location takes into account uncertainty due to lack of knowledge because of insufficient sampling (three factors of Burrough (1986), affect the interpolation variance: density of observations; natural variation and interpolation).

Considering that we have an interpolated raster map and its associated uncertainty, we can perform arithmetic operations with raster maps. Arithmetic map operations are very common procedures used in geographic information system. In general, two raster maps are combined using arithmetic operators (addition, subtraction, multiplication and division) to derive a new raster map. Let X and Y be interpolated raster maps from the same data set and let us perform some arithmetic operation between them. For each variable we know not only the interpolated value but also the uncertainty given by the interpolation variance. Thus, we have to calculate the resulting raster map after an arithmetic operation and the uncertainty as well. Uncertainty is magnified when cartographic overlay operations involve more than two steps (Burrough 1986). We can define a function of variables X and Y: f(x,y) that can be expanded in a Taylor series about the mean values of X and Y. From this expansion we can apply the mathematical expectation operator to find the mean value of the function and the definition of the variance as a measurement of the uncertainty. Addition and subtraction are very simple arithmetic operators, but multiplication and division are much more complex operations and they require correcting the combined estimate and the calculation of the variance is even more complicated.

ESTIMATES AND VARIANCES OF ARITHMETICALLY COMBINED VARIABLES

Given a random variable X, the mathematical expectation is:

E [ X ] = 1 n i = 1 n X i

The variance is a measure of the uncertainty associated with the mathematical expectation:

𝑉 𝑎 𝑟 [ X ] = 1 n i = 1 n ( X i E [ X ] ) 2

which can be written as:

𝑉 𝑎 𝑟 [ X ] = E [ X E [ X ] ] 2 = E [ X 2 ] ( E [ X ] ) 2

Given two random variables (X and Y), we can compute the covariance as:

𝐶 𝑜 𝑣 ( X , Y ) = E [ 𝑋 𝑌 ] E [ X ] E [ Y ]

The correlation coefficient as a measure of the mutual relationship between two random variables is calculated as:

ρ X , Y = 𝐶 𝑜 𝑣 ( X , Y ) 𝑉 𝑎 𝑟 [ X ] 𝑉 𝑎 𝑟 [ Y ] = 𝐶 𝑜 𝑣 ( X , Y ) S X S Y

Where Sx and Sy are standard deviations.

Let us call f(x,y) a function resulting from arithmetically combined random variables X and Y. We are interested in the mean value of the function f(x,y) around the mean values of random variables X and Y. To find the mean and variance of this function we use Taylor expansion around θ, which is the point near the mean values of X and Y: θ=(μx,μy). For the following development let us consider a simplified notation in which: μx=E[X];μy=E[Y];σx2=𝑉𝑎𝑟[X];σy2=𝑉𝑎𝑟[Y] and σ𝑥𝑦=𝐶𝑜𝑣(X,Y). Notice that for propagation of uncertainty we need to know the means, variances, the covariance, and the type of operation (Heuvelink et al. 19897 HEUVELINK GB, BURROUGH PA AND STEIN A. 1989. Propagation of errors in spatial modelling with GIS. Int J Geogr Inf Syst 3: 303-322.).

The second-order Taylor expansion is (Weir and Hass 201414 WEIR MD AND HASS J. 2014. Thomas’ calculus – multivariable. Boston, Person Education Inc, 13th ed., p. 572-1044.):

f ( x , y ) = f ( θ ) + f x ( θ ) ( x μ x ) + f y ( θ ) ( y μ y ) + 1 2 f 𝑥 𝑥 ( θ ) ( x μ x ) 2 + f 𝑥 𝑦 ( θ ) ( x μ x ) ( y μ y ) + 1 2 f 𝑦 𝑦 ( θ ) ( y μ y ) 2 + 𝑟 𝑒 𝑚 𝑎 𝑖 𝑛 𝑑 𝑒 𝑟

Applying the expectation operator in (5) and neglecting the remainder, we have:

E [ f ( x , y ) ] = f ( θ ) + 1 2 f 𝑥 𝑥 ( θ ) σ x 2 + f 𝑥 𝑦 ( θ ) σ 𝑥 𝑦 + 1 2 f 𝑦 𝑦 ( θ ) σ y 2

This is the general expression for computing the mathematical expectation of the function f(x,y) about θ=(μx,μy).

The variance of the function f(x,y) around θ=(μx,μy) can be computed as:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E [ ( f ( x , y ) E ( f ( x , y ) ) ) 2 ]

Replacing definitions (6) and (7) in (3), we have:

𝑉 𝑎 𝑟 [ f ( x , y ) ] = E [ ( f x ( θ ) ( x μ x ) + f y ( θ ) ( y μ y ) + 1 2 f 𝑥 𝑥 ( θ ) ( x μ x ) 2 + f 𝑥 𝑦 ( x μ x ) ( y μ y ) + 1 2 f 𝑦 𝑦 ( θ ) ( y μ y ) 2 1 2 f 𝑥 𝑥 ( θ ) σ x 2 f 𝑥 𝑦 ( θ ) σ 𝑥 𝑦 1 2 f 𝑦 𝑦 ( θ ) σ y 2 ) 2 ]

This expression cannot be simplified because it depends on partial derivatives of the function f(x,y).

The mean and variance of the function f(x,y) are presented in Table I. All formulas result from second-order Taylor expansion (except for equations (13) and (14)) and are equal to those derived by Mood et al. (1974). Equation (13) results from third-order Taylor expansion. This equation should provide a better approach for the mean of quotient of random variables. The variance after equation (14) uses only first-order terms (Mood et al. 197411 MOOD AM, GRAYBILL FA AND BOES DC. 1974. Introduction to the theory of statistics. Tokyo, McGraw Hill, 3rd ed., 564 p.).

TABLE I
Mean and variance for the function f(x,y) around θ=(μx,μy).

It can be noted that for the quotient of X and Y, the mean and variance are calculated by approximate formulas. For a better approximation of the variance of the quotient of X and Y, we considered second-order terms according to Maskey and Guinot (2003). The variance of the quotient after equation (15) is better than the variance computed after equation (14). Indeed, the algebraic complexity increased very much with the inclusion of second-order terms (Maskey and Guinot 20039 MASKEY S AND GUINOT V. 2003. Improved first-order second moment method for uncertainty estimation in flood forecasting. Hydrolog Sci J 48: 183-196.). The mean and variance for the quotient of X and Y will always result in approximate formulas because the function f(x,y)=X/Y is infinitely differentiable with respect to Y. An exact formula for the calculation of the variance of the quotient of X and Y was proposed by Frishman (19714 FRISHMAN F. 1971. On the arithmetic means and variances of products and ratios of random variables. National Technical Information Service, AD-785623, p. 331-345.). However, this alternative uses the variance of the product X*1/Y, which is implied in estimating 1/Y instead of Y. Details of the mathematical development of Taylor expansion to find the mean and variance of the function f(x,y) can be found in the Appendix. This development was based entirely on statistics (Mood et al. 1974) and calculus (Weir and Hass 2014).

Heuvelink et al. (1989) proposed a general formula based on second order Taylor expansion for n variables. The mean of the function f(M) around M={μ1,μ2,,μn} is computed as (Heuvelink et al. 1989):

Where ρ𝑖𝑗 is the correlation coefficient between variables i and j; σi is the standard deviation for variable i and σj is the standard deviation for variable j.

The general formula for variance of the function f(M) is (Heuvelink et al. 1989):

𝑉 𝑎 𝑟 ( x ́ ) = E { [ k = 1 n [ ( x k μ k ) f k ( M ) ] + 1 2 i = 1 n j = 1 n [ ( x i μ i ) ( x j μ j ) ρ 𝑖 𝑗 σ i σ j ? f 𝑖 𝑗 ( M ) ] ] 2 }
𝑉 𝑎 𝑟 ( x ́ ) = k = 1 n l = 1 n ρ 𝑘 𝑙 σ k σ l f k ( M ) f l ( M ) + ? k = 1 n i = 1 n j = 1 n { E [ ( x k μ k ( ( x i μ i ) ( x j μ j ) ρ 𝑘 𝑙 σ k σ l ) ) ] ( f k ( M ) ) * f 𝑖 𝑗 ( M ) } + ? 1 4 i = 1 n j = 1 n k = 1 n l = 1 n { E [ ( x i μ i ) ( x j μ j ) ( x k μ k ) ( x l μ l ) ρ 𝑖 𝑗 σ i σ j ρ 𝑘 𝑙 σ k σ l ] f 𝑖 𝑗 ( M ) f 𝑘 𝑙 ( M ) }

We proved that all formulas presented in Table I (except equations (12) and (14)) are equal to those developed from Heuvelink’s general formulas for mean and variance (Appendix).

LOCAL ESTIMATES AND UNCERTAINTIES USING MULTIQUADRIC EQUATIONS

This paper concerns building raster maps from field data, which means interpolation of regularly spaced cells based on neighbor data. Multiquadric equations (Hardy 19715 HARDY RL. 1971. Multiquadric equations of topography and other irregular surfaces. J Geophys Res 76: 1905-1915.) were chosen for interpolation of the variable at unsampled locations. Usually, several variables are measured in each sample location. For instance, in geochemical exploration, major and trace elements are analyzed simultaneously for each sample. Once we have built raster maps with these variables, we can combine variables to get another raster map. This is a common procedure in GIS. We present a method that allows combination of random variables and also uncertainty quantification based on equations summarized in Table I. It is important to emphasize that it depends on a reliable measure of uncertainty. In this sense, we propose the use of the interpolation variance (Yamamoto 2000) as an approach for uncertainty quantification. Interpolation variance can be used with any estimate based on weighted average formula.

For each cell location within the raster map, we have to find the nearest neighbor data. The dual form of multiquadric interpolation can be written as (Yamamoto 200217 YAMAMOTO JK. 2002. Ore reserve estimation using radial basis functions. Rev Inst Geol 23: 25-38.):

Z 𝑀 𝑄 * ( x o ) = i = 1 n w i Z ( x i )

The weights {wi,i=1,,n} come from the solution of a system of linear equations (Yamamoto 2002):

j = 1 n w j ( x i x j ) + μ = ( x i x o ) 𝑓 𝑜 𝑟 i = 1 j = 1 n w j = 1

Where μ is the Lagrange multiplier and (x)=|x|2+C is the multiquadric kernel as proposed by Hardy (1971).

Because equation (17) is a weighted average formula, we can use the interpolation variance for quantifying the uncertainty associated with multiquadric interpolation. The interpolation variance is simply (Yamamoto 2000):

S o 2 ( x o ) = i = 1 n w i [ Z ( x i ) Z 𝑀 𝑄 * ( x o ) ] 2

Raster map operations consist of combining two variables at once. Let us call Z1(x) and Z2(x) the two variables to be arithmetically combined. Both variables are measured on same data location and the data set has the same number of data points. This is the essential condition to perform arithmetical operation between variables Z1(x) and Z2(x). Since we are using multiquadric equations with the same multiquadric kernel, then the weights are identical: {wi1=wi2,i=1,,n}. In light of this fact we will consider hereafter the set of weights {wi,i=1,,n} for both variables.

Multiquadric interpolations for variables Z1(x) and Z2(x) are given as:

Z 𝑀 𝑄 1 * ( x o ) = i = 1 n w i Z 1 ( x i )
Z 𝑀 𝑄 2 * ( x o ) = i = 1 n w i Z 2 ( x i )

The term E[Z1(x)Z2(x)] as a local statistic can be calculated as:

E [ Z 1 ( x ) Z 2 ( x ) ] = i = 1 n ( w i Z 1 ( x i ) Z 2 ( x i ) )

Thus, the local covariance can be computed as:

𝐶 𝑜 𝑣 ( Z 1 ( x ) , Z 2 ( x ) ) = i = 1 n w i Z 1 ( x i ) Z 2 ( x i ) ( i = 1 n w i Z 1 ( x i ) ) ( i = 1 n w i Z 2 ( x i ) )

See that equation (27) allows computing cell-by-cell covariance. This statistic is fundamental for propagation of uncertainty. The interpolation covariance in (27) is calculated directly from neighbor data using the same weights for multiquadric interpolation. Therefore, this approach is different from the spatial covariance derived from the variogram model. Actually, for Z1(x)=Z2(x), this formula is equivalent to the interpolation variance.

Interpolation variances are computed as:

S o 1 2 = i = 1 n w i [ Z 1 ( x i ) Z 𝑀 𝑄 * ( x o ) ] 2
S o 2 2 = i = 1 n w i [ Z 2 ( x ? i ) Z 𝑀 𝑄 * ( x o ) ] 2

Means and variances in Table I concern global statistics. But, raster maps imply local statistics and the purpose is to show that local statistics can be combined to compute mean and variances after mathematical operations between two random variables.

Keeping the same notation as used in Table I, we consider the following equivalence: μx=Z𝑀𝑄1*(x);μy=Z𝑀𝑄2*(x);σx2=So12(xo);σy2=So22(xo) and σ𝑥𝑦=𝐶𝑜𝑣(Z1(x),Z2(x)).

In addition, high-order moments in Table I are calculated as:

E [ . ] = i = 1 n w i ( . )

MATERIALS AND METHODS

For this study we consider a stratified random sample drawn from an exhaustive data set composed of 2500 data points arranged in an array of 50 rows by 50 columns. This sample, with 100 data points (4% of the exhaustive) presents two measured variables (VP and VS2) in each location (Figure 1a). These variables present a positive correlation (Figure 1b). In addition to performing arithmetic operations between VP and VS2, we have to prove that mean and variances of combined variables are valid. Thus, we generated four new variables: VP+VS2; VP-VS2; VPxVS2 and VP/VS2. Sample statistics for all variables are presented in Table II.

Figure 1
Location map of sample data (a) and VP X VS2 scattergram (b) (see the colors in the online version).
TABLE II
Sample statistics for variables VP, VS2, VP+VS2, VP-VS2, VPxVS2 and VP/VS2.

VP and VS2 will be used for building raster maps, from which we want to obtain new raster maps by combining them arithmetically. Mean and variance for new random variables are listed in Table III. Comparing these results with directly computed statistics in Table II, we verify that except for the quotient of random variables, all others are exactly the same.

The sample data with six variables are the materials of this study. Original variables VP and VS2 will be used to build raster maps based on multiquadric interpolation. Equations (24) and (28) will be used to compute cells of the VP raster maps and equations (25) and (29) for VS2 raster maps. The covariance between VP and VS2 will be calculated after equation (27). These statistics will be combined according to formulas in Table I to derive means and uncertainties of resulting arithmetically combined raster maps. Now, we can compare resulting raster maps with those computed directly from new variables based on equations (16) and (17). Next, these equations ((21) and (23)) will be checked against equations (10) and (11) for variables VP±VS2, equations (12) and (13) for variable VPxVS2 and equations (14), (15), (16) and (17) for variable VP/VS2.

TABLE III
Mean and variance for combined random variables.

RESULTS AND DISCUSSION

Raster maps for variables VP and VS2 are presented in Figure 2. For the purpose of displaying interpolation variances and covariance, we considered a maximum equal to 95% of their distributions. Now, raster maps for VP and VS2 can be combined by applying arithmetic operators. Before proceeding it is important to show raster maps (Figure 3) computed directly from new variables.

The first operation is the sum: VP+VS2. Results of this operation are presented in Figure 4. The result (Figure 4a) is simply the sum of raster maps (for VP and VS2). The variance of the sum (Figure 4b) is equal to the sum of variances for VP (σx2) and VS2 (σy2) plus twice the covariance between VP and VS2 (σ𝑥𝑦). The variance of the sum has great influence of the variable VP and the covariance as well. Results illustrated in Figure 4 are exactly the same as shown in Figure 3a.

Results of the subtraction of VS2 from VP are presented in Figure 5. The mathematical expectation and variance follow the same raster operations as for the sum. However, the variance is equal to the sum of variances minus twice the covariance between VP and VS2. The variance of the difference between random variables VP and VS2 shows a reasonable correlation with the variance of VP, but almost no correlation with the variance of VS2 and the covariance. Mean and variances computed after equations (10) and (11) for addition and subtraction match those values computed directly from equations (21) and (23). Once again, results can be compared graphically with images presented in Figure 3b.

Figure 2
Raster maps for variables VP (a) and for variable VS2 (b). Interpolated variables in column I, uncertainties in column II and covariance between VP and VS2 in column III. Interpolation variances and covariance were displayed for a maximum equal to 95% of their distributions (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).
Figure 3
Raster maps computed directly from random variables VP+VS2 (a); VP-VS2 (b); VPxVS2 (c) and VP/VS2 (d). I) Interpolated variables; II) variances (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).
Figure 4
Results of arithmetic operation VP+VS2. The result (a) is equal to the sum of VP (μx) and VS2 (μy). The associated uncertainty (b) is equal to the variance of VP (σx2) plus variance of VS2 (σy2) plus twice the covariance between VP and VS2 (σ𝑥𝑦).
Figure 5
Results of arithmetic operation VP-VS2. The result (a) is equal to the sum of VP (μx) and VS2 (μy). The associated uncertainty (b) is equal to the variance of VP (σx2) plus variance of VS2 (σy2) minus twice the covariance between VP and VS2 (σ𝑥𝑦) (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).

The results of the multiplication of VP and VS2 are presented in Figure 6, which can be compared with results of direct multiquadric interpolation of the variable VPxVS2 (Figure 3c). The mathematical expectation of this operation E[f(x,y)] is equal to the product of VP and VS2 plus the covariance between VP and VS2. Note that the mathematical expectation has the contribution of the covariance. E[f(x,y)] is highly correlated with the means μx and μy and low correlation with the covariance. The uncertainty 𝑉𝑎𝑟[f(x,y)] is equal to term 1 (μy2σx2) plus term 2 (μx2σy2) plus term 3 (E[(xμx)2(yμy)2]) minus term 4 (σ𝑥𝑦2) plus term 5 (2μxμyσ𝑥𝑦) plus term 6 (2μyE[(xμx)2(yμy)]) plus term 7 (2μxE[(xμx)(yμy)2]). Next, the variance was calculated after second-order Taylor expansion (equation (13)) and therefore higher order moments have been included. Equations (12) and (13) provided the same results as applying equations (21) and (23) for the new random variable VPxVS2.

Figure 6
Results of arithmetic operation VPxVS2. The result (E[f(x,y)]) is equal to the product of VP (μx) and VS2 (μy) plus the covariance between VP and VS2 (σ𝑥𝑦). The associated uncertainty (𝑉𝑎𝑟[f(x,y)]) is equal to the term 1 (μy2σx2) plus the term 2 (μx2σy2) plus the term 3 (E[(xμx)2(yμy)2]) minus the term 4 (σ𝑥𝑦2) plus the term 5 (2μxμyσ𝑥𝑦) plus the term 6 (2μyE[(xμx)2(yμy)]) plus the term 7 (2μyE[(xμx)(yμy)2]) (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).

The calculation of mean and variance for the quotient of VP and VS2 is more complex since the function X/Y is infinitely differentiable with respect to Y. The mathematical expectation is not only equal to the quotient between VP and VS2 (μx/μy), but includes two other terms: σ𝑥𝑦/μy2 and μxσy2/μy3(Figure 7a). This is the result after equation (14), which results from second-order Taylor expansion. Results from application of the equation (15) are shown in Figure 7b. Although results look very similar to each other, the mean according to equation (15) is more accurate than that provided by equation (14). Figures 7a, b basically display the same pattern as shown in Figure 3d-I. For calculating the variance of the ratio of X to Y, we have equation (16) that results from first-order Taylor expansion and another equation (17) derived after second-order Taylor expansion (Figure 8). It is virtually impossible to see the difference between results of these approaches in Figure 8. Thus, means and variances can be compared by being displayed in a scatter plot (Figure 9). Although, mean values are very close to each other, we can see that equation (15) provides a better approach than the equation (14). On the other hand, the variance after equation (17) is much better than that provided by equation (16).

We can follow arithmetic operations taking as an example the cell with coordinates CX=29.50 and CY=23.5 (Table IV). With eight neighbor data we interpolated this cell using field data for VP and VS2. All other variables (VP+VS2, VP-VS2, VPxVS2 and VP/VS2) were derived from these measurements (VP and VS2). Note that we can compute mean and variance directly for derived variables and thus these values can be checked against statistics computed using formulas listed in Table I. In Table IV, the data point with coordinates (CX=26.5, CY=25.5) has a weight equal to zero. This is because we applied an algorithm for correcting negative weights (Rao and Journel 199712 RAO SE AND JOURNEL AG. 1997. Deriving conditional distributions from ordinary kriging. In: Baafi EY and Schofield NA (Eds), Geostatistics Wollongong 96: 92-102.), in which a constant equal to the modulus of the largest negative weight was added to all weights and then restandardized to a sum equal to one.

Figure 7
Mean values for arithmetic operation VP/VS2. (a - upper row) mean after equation (14); (b - lower row) mean according to equation (15) (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).
Figure 8
Variances of the arithmetic operation VP/VS2. a) resulting variance from first order Taylor expansion (equation (15)); b) variance calculated using second order Taylor expansion (equation (17)) (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).
Figure 9
Scatterplots for means of the ratio of X to Y: a = equation (14) and b = equation (15); variances of the ratio: c = equation (16) and d = equation (17) (for the interpretation of the references to color in this figure legend, the reader is referred to the web version of this article).
TABLE IV
Neighboring data around a cell with coordinates (CX=29.50, CY=23.5), mean and variances for variables VP, VS2, VP+VS2, VP-VS2, VPxVS2 and VP/VS2.

Mean and variances for the cell with coordinates (CX=29.50, CY=23.5) after combining random variables VP and VS2 are presented in Table V. As can be seen, statistics for VP+VS2, VP-VS2 and VPxVS2 match with statistics computed directly from derived variables listed in Table IV. However, for VP/VS2 we obtain only approximate results.

TABLE V
Mean and variance for combined random variables.

CONCLUSIONS

Propagation of uncertainty coming from arithmetic operations between random variables is well known in statistics (Mood et al. 1974). Heuvelink et al. (1989) and Heuvelink (19986 HEUVELINK GB. 1998. Error propagation in environmental modelling with GIS. London, Taylor & Francis, 127 p.) have established the method of calculating mean and variances of the output raster map from several input maps based on Taylor method. This paper showed that propagation of uncertainty depends on a reliable measure of local accuracy and local covariance. We proved that it is possible to get exact mean and variance of the output raster map for map operations involving addition, subtraction and multiplication. For the mean of quotient, we proposed to compute it from a third-order Taylor expansion that provides a slightly better result than the mean computed from second-order Taylor expansion. For the variance of the ratio of X to Y, we demonstrated that the formula based on second-order Taylor expansion provides much better results than the first order.

ACKNOWLEDGMENTS

The authors wish to thank Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq) for supporting this research.

  • 1
    ATKINSON PM AND FOODY GM. 2002. Uncertainty in remote sensing and GIS: fundamentals. In: Foody GM and Atkinson PM (Eds), Uncertainty in Remote Sensing and GIS, p. 1-18.
  • 2
    BURROUGH PA. 1986. Principles of geographical information system. Oxford, Oxford University Press, 194 p.
  • 3
    CROSETTO M, TARANTOLA S AND SALTELI A. 2000. Sensitivity and uncertainty analysis in spatial modelling based on GIS. Agr Ecosyst Environ 81: 71-79.
  • 4
    FRISHMAN F. 1971. On the arithmetic means and variances of products and ratios of random variables. National Technical Information Service, AD-785623, p. 331-345.
  • 5
    HARDY RL. 1971. Multiquadric equations of topography and other irregular surfaces. J Geophys Res 76: 1905-1915.
  • 6
    HEUVELINK GB. 1998. Error propagation in environmental modelling with GIS. London, Taylor & Francis, 127 p.
  • 7
    HEUVELINK GB, BURROUGH PA AND STEIN A. 1989. Propagation of errors in spatial modelling with GIS. Int J Geogr Inf Syst 3: 303-322.
  • 8
    JOURNEL AG AND ROSSI M. 1989. When do we need a trend model in kriging? Math Geolo 21: 715-739.
  • 9
    MASKEY S AND GUINOT V. 2003. Improved first-order second moment method for uncertainty estimation in flood forecasting. Hydrolog Sci J 48: 183-196.
  • 10
    MCCALLUM WG. 1998. Multivariable calculus. New York, J Wiley & Sons, 270 p.
  • 11
    MOOD AM, GRAYBILL FA AND BOES DC. 1974. Introduction to the theory of statistics. Tokyo, McGraw Hill, 3rd ed., 564 p.
  • 12
    RAO SE AND JOURNEL AG. 1997. Deriving conditional distributions from ordinary kriging. In: Baafi EY and Schofield NA (Eds), Geostatistics Wollongong 96: 92-102.
  • 13
    WANG G, GERTNER GZ, FANG S AND ANDERSON AB. 2005. A methodology for spatial uncertainty analysis of remote sensing and GIS products Photogramm Eng Rem S 71: 1423-1432.
  • 14
    WEIR MD AND HASS J. 2014. Thomas’ calculus – multivariable. Boston, Person Education Inc, 13th ed., p. 572-1044.
  • 15
    WELLMANN FJ, HOROWITZ GF, SCHILL E AND REGENAUER-LIEB K. 2010. Towards incorporating uncertainty of structural data in 3D geological inversion. Tectonophysics 490: 141-152.
  • 16
    YAMAMOTO JK. 2000. An alternative measure of the reliability of ordinary kriging estimates. Math Geol 32: 489-509.
  • 17
    YAMAMOTO JK. 2002. Ore reserve estimation using radial basis functions. Rev Inst Geol 23: 25-38.
  • 18
    YAMAMOTO JK, MAO XM, KOIKE K, CROSTA AP, LANDIM PMB, HU HZ, WANG CY AND YAO LQ. 2012. Mapping an uncertainty zone between interpolated types of a categorical variable. Comput Geosci 40: 146-152.

APPENDIX

Mean and variance of the function f(x,y) using Taylor expansion

In this appendix, we develop equations for mean and variance of the function f(x,y) as presented in Table I. We will use Taylor expansion of f(x,y) to find the mean and variance of this function around mean values μx and μy. Besides mean values, variances of random variables (σx2 and σy2) and the covariance between random variables X and Y (σ𝑥𝑦) must be known. The development for the mean and variance of product and quotient of random variables was based on Taylor expansion, according to Mood et al. (1974).

According to Weir and Hass (2014), the general equation of Taylor’s Formula for two variables at the point θ=(a,b) is:

Where h=(xa), k=(yb), 0<c<1 and the last term is the remainder.

Neglecting the remainder term and taking n=2, we have:

f ( x , y ) = f ( a , b ) + f ( a , b ) x ( x a ) + f ( a , b ) y ( y b ) + 1 2 ! ( 2 f ( a , b ) x 2 ( x a ) 2 + 2 2 f ( a , b ) x y ( x a ) ( y b ) + 2 f ( a , b ) y 2 ( y b ) 2 )

Letting (a,b)=(μx,μy) in equation A1.2 we have:

f ( x , y ) = f ( μ x , μ y ) + f x ( μ x , μ y ) ( x μ x ) + f y ( μ x , μ y ) ( y μ y ) + 1 2 ! ( f 𝑥 𝑥 ( μ x , μ y ) ( x μ x ) 2 + 2 f 𝑥 𝑦 ( μ x , μ y ) ( x μ x ) ( y μ y ) + f 𝑦 𝑦 ( μ x , μ y ) ( y μ y ) 2 )

Applying the expectation operator to expression A1.3:

E [ f ( x , y ) ] = E [ f ( μ x , μ y ) ] + E [ f x ( μ x , μ y ) ( x μ x ) ] + E [ f y ( μ x , μ y ) ( y μ y ) ] + 1 2 ! ( E [ f 𝑥 𝑥 ( μ x , μ y ) ( x μ x ) 2 ] + 2 E [ f 𝑥 𝑦 ( μ x , μ y ) ( x μ x ) ( y μ y ) ] + E [ f 𝑦 𝑦 ( μ x , μ y ) ( y μ y ) 2 ] )

Finally:

E [ f ( x , y ) ] = f ( μ x , μ y ) + 1 2 ( σ x 2 f 𝑥 𝑥 ( μ x , μ y ) + 2 σ 𝑥 𝑦 f 𝑥 𝑦 ( μ x , μ y ) + σ y 2 f 𝑦 𝑦 ( μ x , μ y ) )

This is the general expression for mathematical expectation of f(x,y). The variance of the function f(x,y) can be written as:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E .

Replacing A1.2 and A1.3 in A1.4, we get:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E .

Calculation of mean and variance for sum and subtraction of random variables X and Y

For f(x,y)=x±y, we have the following derivatives:

f x ( x , y ) = 1 f y ( x , y ) = ± 1 f 𝑥 𝑥 ( x , y ) = 0 f 𝑥 𝑦 ( x , y ) = 0 f 𝑦 𝑦 ( x , y ) = 0

Applying them on equations A1.4 and A1.6, we have equations (10) and (11) of Table I:

E [ f ( x , y ) ] = f ( μ x , μ y )
E [ f ( x , y ) ] = μ x ± μ y
𝑉 𝑎 𝑟 ( f ( x , y ) ) = E [ ( ( x μ x ) ± ( y μ y ) ) 2 ]
𝑉 𝑎 𝑟 ( f ( x , y ) ) = E [ ( x μ x ) 2 ± 2 ( x μ x ) ( y μ y ) + ( y μ y ) 2 ]
𝑉 𝑎 𝑟 ( f ( x , y ) ) = σ x 2 + σ y 2 ± 2 σ 𝑥 𝑦

Calculation of the mean and variance for the product of random variables X and Y

For f(x,y)=𝑥𝑦, we have the following derivatives:

fx(x,y)=yfy(x,y)=xf𝑥𝑥(x,y)=0f𝑦𝑦(x,y)=0f𝑥𝑦(x,y)=1
Applying them on equation A1.4, we get equation ((12) - Table I):

E [ f ( x , y ) ] = μ x μ y + σ 𝑥 𝑦

Replacing derivatives in equation A1.6, we have:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E { [ μ y ( x μ x ) + μ x ( y μ y ) + ( x μ x ) ( y μ y ) σ 𝑥 𝑦 ] 2 }

Notice that:

( a + b + c + d ) 2 = a 2 + b 2 + c 2 + d 2 + 2 𝑎 𝑏 + 2 𝑎 𝑐 + 2 𝑎 𝑑 + 2 𝑏 𝑐 + 2 𝑏 𝑑 + 2 𝑐 𝑑

Expanding A1.10 according to A1.11 we obtain:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E .

Applying definitions of variance, covariance and the general equation of variance A1.6, we get equation ((13) - Table I):

𝑉 𝑎 𝑟 ( f ( x , y ) ) = μ y 2 σ x 2 + μ x 2 σ y 2 + 2 μ x μ y σ 𝑥 𝑦 + E [ ( x μ x ) 2 ( y μ y ) 2 ] ( σ 𝑥 𝑦 ) 2 + 2 μ y E [ ( x μ x ) 2 ( y μ y ) ] + 2 μ x E [ ( x μ x ) ( y μ y ) 2 ]

Calculation of the mean and variance for the ratio of X to Y

For calculation of the mean of the quotient of random variables X and Y we will consider second and third order Taylor expansion as follows.

For function f(x,y)=x/y, we have the following derivatives:

f x = 1 y f y = x y 2 f 𝑥 𝑥 = 0 f 𝑥 𝑦 = 1 y 2 f 𝑦 𝑦 = 2 x y 3 f 𝑥 𝑥 𝑥 = 0 f 𝑥 𝑥 𝑦 = 0 f 𝑥 𝑦 𝑦 = 2 y 3 f 𝑦 𝑦 𝑦 = 6 x y 4

Second order Taylor expansion for calculation of the mean of function f(x,y)=x/y:

Applying the derivatives (until second order) on the equation A1.4 we get (equation (14) - Table I):

E [ f ( x , y ) ] = μ x μ y σ 𝑥 𝑦 2 1 μ y 2 + σ y 2 μ x μ y 3

Third order Taylor expansion for calculation of the mean of function f(x,y)=x/y:

Developing equation A1.1 for n=3, we get:

f ( x , y ) = f ( a , b ) + f ( a , b ) x ( x a ) + f ( a , b ) y ( y b ) + 1 2 ! ( 2 f ( a , b ) x 2 ( x a ) 2 + 2 2 f ( a , b ) x y ( x a ) ( y b ) + 2 f ( a , b ) y 2 ( y b ) 2 ) + 1 3 ! ( 3 f ( a , b ) x 3 ( x a ) 3 + 3 3 f ( a , b ) x 2 y ( x a ) 2 ( y b ) + 3 3 f ( a , b ) x y 2 ( x a ) ( y b ) 2 + 3 f ( a , b ) y 3 ( y b ) 3 )

Letting (a,b)=(μx,μy) in equation A1.14, we have:

f ( x , y ) = f ( μ x , μ y ) + f x ( μ x , μ y ) ( x μ x ) + f y ( μ x , μ y ) ( y μ y ) + 1 2 ! ( f 𝑥 𝑥 ( μ x , μ y ) ( x μ x ) 2 + 2 f 𝑥 𝑦 ( μ x , μ y ) ( x μ x ) ( y μ y ) + f 𝑦 𝑦 ( μ x , μ y ) ( y μ y ) 2 ) + 1 3 ! ( f 𝑥 𝑥 𝑥 ( μ x , μ y ) ( x μ x ) 3 + 3 f 𝑥 𝑥 𝑦 ( μ x , μ y ) ( x μ x ) 2 ( y μ y ) + 3 f 𝑥 𝑦 𝑦 ( μ x , μ y ) ( x μ x ) ( y μ y ) 2 + f 𝑦 𝑦 𝑦 ( y μ y ) 3 )

Replacing derivatives in equation A1.15 we have (equation (15) - Table I):

E [ f ( x , y ) ] = μ x μ y σ 𝑥 𝑦 1 μ y 2 + σ y 2 μ x μ y 3 + E [ ( x μ x ) ( y μ y ) 2 ] μ y 3 μ x μ y 4 E [ ( y μ y ) 3 ]

First order Taylor expansion for calculation of the variance of function f(x,y)=x/y:

To obtain the expression for variance using First order Taylor expansion, we will develop the equation A1.1 without the summatory term, then we have:

f ( x , y ) = f ( a , b ) + f ( a , b ) x ( x a ) + f ( a , b ) y ( y b )

Letting (a,b)=(μx,μy), we have:

f ( x , y ) = f ( μ x , μ y ) + f x ( μ x , μ y ) ( x μ x ) + f y ( μ x , μ y ) ( y μ y )

Then, we will get for the expectation:

E [ f ( x , y ) ] = f ( μ x , μ y )

And for the variance:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E [ ( f x ( μ x , μ y ) ( x μ x ) + f y ( μ x , μ y ) ( y μ y ) ) 2 ]
𝑉 𝑎 𝑟 ( f ( x , y ) ) = E [ f x 2 ( μ x , μ y ) ( x μ x ) 2 + 2 f x ( μ x , μ y ) f y ( μ x , μ y ) ( x μ x ) ( y μ y ) + f y 2 ( μ x , μ y ) ( y μ y ) 2 ]

Applying the corresponding derivatives (until first order) we get (equation (16) - Table I):

𝑉 𝑎 𝑟 ( f ( x , y ) ) = σ x 2 μ y 2 2 μ x μ 3 σ 𝑥 𝑦 + μ x 2 μ y 4 σ y 2

Second order Taylor expansion for calculation of the variance of function f(x,y)=x/y (equation A1.6):

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E { [ 1 μ y ( x μ x ) μ x μ y 2 ( y μ y ) 1 μ y 2 [ ( x μ x ) ( y μ y ) σ 𝑥 𝑦 ] + μ x μ y 3 [ ( y μ y ) 2 σ y 2 ] ] 2 }
𝑉 𝑎 𝑟 ( f ( x , y ) ) = E { [ ( x μ x ) μ y μ x μ y 2 ( y μ y ) + [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] μ y 2 + μ x μ y 3 [ ( y μ y ) 2 σ y 2 ] ] 2 }

Expanding A1.19 according to A1.11 we obtain:

𝑉 𝑎 𝑟 ( f ( x , y ) ) = E .

Developing and simplifying, we get equation ((17) - Table I):

𝑉 𝑎 𝑟 [ f ( x , y ) ] σ x 2 μ y 2 + μ x 2 μ y 4 σ y 2 2 μ x μ y 3 σ 𝑥 𝑦 + 1 μ y 4 E [ ( σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ) 2 ] + μ x 2 μ y 6 E [ ( ( y μ y ) 2 σ y 2 ) 2 ] + 2 μ y 3 E [ ( x μ x ) [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] ] + 2 μ x μ y 4 E [ ( x μ x ) [ ( y μ y ) 2 σ y 2 ] ] 2 μ x μ y 4 E [ ( y μ y ) [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] ] 2 μ x 2 μ y 5 E [ ( y μ y ) [ ( y μ y ) 2 σ y 2 ] ] + 2 μ x μ y 5 E [ [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] [ ( y μ y ) 2 σ y 2 ] ]
Heuvelink’s approach for calculation of the mean and variance of function f(x,y)based on second order Taylor expansion

In this appendix we wanted to show that the traditional approach used in Appendix A1 gives the same results as obtained by Heuvelink et al. (1989) and Heuvelink (1998). Actually, these references presented a general formula of second order Taylor expansion for n random variables.

The second order Taylor expansion of a function with n variables at the point M=(µ1,...,µn) and neglecting the remainder can be written as (Heuvelink et al. 1989):

f ( x ́ ) = f ( M ) + i = 1 n [ ( x i μ i ) f x i ( M ) ] + 1 2 i = 1 n j = 1 n [ ( x i μ i ) ( x j μ j ) f x i x j ( M ) ]

Then, the mean around M is computed as (Heuvelink 1989):

E ( f ( M ) ) = E [ f ( M ) ] + E [ i = 1 n [ ( x i μ i ) f i ( M ) ] ] + 1 2 E [ i = 1 n j = 1 n [ ( x i μ i ) ( x j μ j ) f 𝑖 𝑗 ( M ) ] ]
E ( f ( M ) ) = f ( M ) + 1 2 i = 1 n j = 1 n [ ρ 𝑖 𝑗 σ i σ j f 𝑖 𝑗 ( M ) ]

Where ρ𝑖𝑗 is the correlation coefficient between variables i and j; σi is the standard deviation for variable i and σj is the standard deviation for variable j.

𝑉 𝑎 𝑟 ( x ́ ) = E { [ k = 1 n [ ( x k μ k ) f k ( M ) ] + 1 2 i = 1 n j = 1 n [ ( ( x i μ i ) ( x j μ j ) ρ 𝑖 𝑗 σ i σ j ) f 𝑖 𝑗 ( M ) ] ] 2 }
𝑉 𝑎 𝑟 ( x ́ ) = k = 1 n l = 1 n ρ 𝑘 𝑙 σ k σ l f k ( M ) f l ( M ) + k = 1 n i = 1 n j = 1 n { E [ ( x k μ k ) ( ( x i μ i ) ( x j μ j ) ρ 𝑘 𝑙 σ k σ l ) ] f k ( M ) f 𝑖 𝑗 ( M ) } + ? 1 4 i = 1 n j = 1 n k = 1 n l = 1 n { E [ ( x i μ i ) ( x j μ j ) ( x k μ k ) ( x l μ l ) ρ 𝑖 𝑗 σ i σ j ρ 𝑘 𝑙 σ k σ l ] * ? * f 𝑖 𝑗 ( M ) f 𝑘 𝑙 ( M ) }

Notice that for n=2, and setting x1=x and x2=y, the expression A2.2 becomes:

E ( f ( x , y ) ) = f ( μ x , μ y ) + 1 2 ( σ x 2 f 𝑥 𝑥 ( μ x , μ y ) + 2 σ 𝑥 𝑦 f 𝑥 𝑦 ( μ x , μ y ) + σ y 2 f 𝑦 𝑦 ( μ x , μ y ) )

Expressions A1.3 and A2.4 are the same, so the development of the expectation for an arithmetic operation is the same as shown in Appendix A1. Therefore, equations (10), (12) and (14) of Table I can be derived from A2.4.

For the variance, we will need to expand the expression A2.3 and consider the arithmetic operation regarding the function f(x,y). Then at the point μ=(μx,μy) we have:

𝑉 𝑎 𝑟 ( x , y ) = σ x 2 f x 2 ( μ ) + 2 σ 𝑥 𝑦 f x ( μ ) f y ( μ ) + σ y 2 f y 2 ( μ ) + E [ ( x μ x ) 3 ( x μ x ) σ x 2 ] f x ( μ ) f 𝑥 𝑥 ( μ ) + 2 E [ ( x μ x ) 2 ( y μ y ) ( x μ x ) σ 𝑥 𝑦 ] f x ( μ ) f 𝑥 𝑦 ( μ ) + E [ ( x μ x ) ( y μ y ) 2 ( x μ x ) σ y 2 ] f x ( μ ) f 𝑦 𝑦 ( μ ) + E [ ( x μ x ) 2 ( y μ y ) ( y μ y ) σ x 2 ] f y ( μ ) f 𝑥 𝑥 ( μ ) + 2 E [ ( x μ x ) ( y μ y ) 2 ( y μ y ) σ 𝑥 𝑦 ] f y ( μ ) f 𝑥 𝑦 ( μ ) + E [ ( y μ y ) 3 ( y μ y ) σ y 2 ] f y ( μ ) f 𝑦 𝑦 ( μ ) + 1 4 .

Calculation of the variance for sum and subtraction of random variables X and Y

Since for f(x,y)=x±y all second order derivatives are equal to zero, we have (equation (11) - Table I):

𝑉 𝑎 𝑟 ( x , y ) = σ x 2 + σ y 2 ± 2 σ 𝑥 𝑦

Because all second order derivatives are equal to zero.

Calculation of the variance for the product of random variables X and Y

𝑉 𝑎 𝑟 ( x , y ) = σ x 2 μ y 2 + σ y 2 μ x 2 + 2 σ 𝑥 𝑦 μ x μ y + 2 μ y E [ ( x μ x ) 2 ( y μ y ) ] + 2 μ x E [ ( x μ x ) ( y μ y ) 2 ] + E [ ( x μ x ) 2 ( y μ y ) 2 ] ( σ 𝑥 𝑦 ) 2

This is the same as equation ((13) - Table I).

Calculation of the variance for the ratio of X to Y

𝑉 𝑎 𝑟 ( x , y ) = σ x 2 μ y 2 2 μ x μ y 3 σ 𝑥 𝑦 + μ x 2 μ y 4 σ y 2 2 μ y 3 E [ ( x μ x ) 2 ( y μ y ) ( x μ x ) σ 𝑥 𝑦 ] + 2 μ x μ y 4 E [ ( x μ x ) ( y μ y ) 2 ( x μ x ) σ y 2 ] + 2 μ x μ y 4 E [ ( x μ x ) ( y μ y ) 2 ( y μ y ) σ 𝑥 𝑦 ] 2 μ x 2 μ y 5 E [ ( y μ y ) 3 ( y μ y ) σ y 2 ] + 1 4 { 4 2 μ x μ y 5 E [ ( x μ x ) ( y μ y ) 3 σ y 2 σ 𝑥 𝑦 ] + 4 1 μ y 4 E [ ( x μ x ) 2 ( y μ y ) 2 ( σ 𝑥 𝑦 ) 2 ] + 4 μ x 2 μ y 6 E [ ( y μ y ) 4 ( σ y 2 ) 2 ] }
𝑉 𝑎 𝑟 ( x , y ) = σ x 2 μ y 2 + μ x 2 μ y 4 σ y 2 2 μ x μ y 3 σ 𝑥 𝑦 + 2 μ y 3 E [ ( x μ x ) [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] ] + 2 μ x μ y 4 E [ ( x μ x ) [ ( y μ y ) 2 σ y 2 ] ] 2 μ x μ y 4 E [ ( y μ y ) [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] ] 2 μ x 2 μ y 5 E [ ( y μ y ) [ ( y μ y ) 2 σ y 2 ] ] 2 μ x μ y 5 E [ ( x μ x ) ( y μ y ) 3 σ y 2 σ 𝑥 𝑦 ] + E [ ( x μ x ) 2 ( y μ y ) 2 ] ( σ 𝑥 𝑦 ) 2 μ y 4 + μ x 2 μ y 6 E [ ( y μ y ) 4 ( σ y 2 ) 2 ]

As we can see, some terms of equation A2.7 do not match A1.21. Actually, we can rewrite these terms as follows.

E [ [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] 2 ] = ( σ 𝑥 𝑦 ) 2 2 σ 𝑥 𝑦 E [ ( x μ x ) ( y μ y ) ] + E [ ( x μ x ) 2 ( y μ y ) 2 ] = E [ ( x μ x ) 2 ( y μ y ) 2 ] ( σ 𝑥 𝑦 ) 2
E [ [ ( y μ y ) 2 σ y 2 ] 2 ] = E [ ( y μ y ) 4 ] 2 σ y 2 E [ ( y μ y ) 2 ] + ( σ y 2 ) 2 = E [ ( y μ y ) 4 ] ( σ y 2 ) 2 = E [ ( y μ y ) 4 ( σ y 2 ) 2 ]
E [ [ σ 𝑥 𝑦 ( x μ x ) ( y μ y ) ] [ ( y μ y ) 2 σ y 2 ] ] = E [ [ ( x μ x ) ( y μ y ) σ 𝑥 𝑦 ] [ σ y 2 ( y μ y ) 2 ] ] = E [ ( x μ x ) ( y μ y ) σ y 2 ] E [ ( x μ x ) ( y μ y ) 3 ] E [ σ 𝑥 𝑦 σ y 2 ] + E [ σ 𝑥 𝑦 ( y μ y ) 2 ] = E [ σ 𝑥 𝑦 σ y 2 ] E [ ( x μ x ) ( y μ y ) 3 ] = E [ σ 𝑥 𝑦 σ y 2 ( x μ x ) ( y μ y ) 3 ]

Replacing these terms into equation A2.7 we get exactly the same as A1.21, that is the equation ((17) - Table I).

Publication Dates

  • Publication in this collection
    Aug 2018

History

  • Received
    03 Aug 2017
  • Accepted
    09 Oct 2017
Academia Brasileira de Ciências Rua Anfilófio de Carvalho, 29, 3º andar, 20030-060 Rio de Janeiro RJ Brasil, Tel: +55 21 3907-8100 - Rio de Janeiro - RJ - Brazil
E-mail: aabc@abc.org.br