2.4.6 Sums of Random Variables



next up previous
Next: 2.4.7 Monte Carlo Integration Up: 2.4 Examples of Continuous Previous: 2.4.5 Bivariate cdf

2.4.6 Sums of Random Variables

Now let us draw samples from the pdf and define the following linear combination of the samples:

 

where the parameters are real constants and the are real-valued functions. Since the are r.v.'s, and is a linear combination of functions of the r.v.'s, is also a r.v. We now examine the properties of , in particular its expectation value and variance. Referring to our earlier discussion of the mean and variance of a linear combination, expressed as Eq. (9) and Eq. (15), respectively, we find

 

 

Now consider the special case where and :

 

Note that is simply the average value of the sampled r.v.'s. Now consider the expectation value for , using Eq. (61):

 

In other words, the expectation value for the average (not the average itself!) of observations of the r.v. is simply the expectation value for . This statement is not as trivial as it may seem, because we may not know in general, because is a property of and the pdf . However, Eq. (62) assures us that an average of observations of will be a reasonable estimate of . Later, we will introduce the concept of an unbiased estimator, and suffice to say for now, that Eq. (62) proves that the simple average is an unbiased estimator for the mean. Now let us consider the variance in , in particular its dependence on the sample size.

Considering again the case where and , and using Eq. (60), the variance in the linear combination is given by:

 

Hence the variance in the average value of samples of is a factor of smaller than the variance in the original r.v. . Note that we have yet to say anything about how to estimate , only that its value decreases as .

This point deserves further elaboration. The quantities and are properties of the pdf and the real function . As mentioned earlier, they are known as the true mean and true variance, respectively, because they are known a priori, given the pdf and the function . Then if we consider a simple average of samples of , denoted , Eq. (62) tells us that the true mean for is equal to the true mean for . On the other hand, Eq. (63) tells us that the true variance for is smaller than the true variance for , an important consequence for estimating errors.

Later we will show how to estimate , an important task since in general we don't know the true mean and variance, and these terms will have to be estimated. Let us now apply this discussion to an important application of Monte Carlo methods, the evaluation of definite integrals.



next up previous
Next: 2.4.7 Monte Carlo Integration Up: 2.4 Examples of Continuous Previous: 2.4.5 Bivariate cdf



verena@csep1.phy.ornl.gov