Wald's Equation

In probability theory, Wald's equation, Wald's identity or Wald's lemma is an important identity that simplifies the calculation of the expected value of the sum of a random number of random quantities.

In its simplest form, it relates the expectation of a sum of randomly many finite-mean, independent and identically distributed random variables to the expected number of terms in the sum and the random variables' common expectation under the condition that the number of terms in the sum is independent of the summands.

The equation is named after the mathematician Abraham Wald. An identity for the second moment is given by the Blackwell–Girshick equation.

Basic version

Let (Xn)nWald's Equation  be a sequence of real-valued, independent and identically distributed random variables and let N ≥ 0 be an integer-valued random variable that is independent of the sequence (Xn)nWald's Equation . Suppose that N and the Xn have finite expectations. Then

    Wald's Equation 

Example

Roll a six-sided dice. Take the number on the die (call it N) and roll that number of six-sided dice to get the numbers X1, . . . , XN, and add up their values. By Wald's equation, the resulting value on average is

    Wald's Equation 

General version

Let (Xn)nWald's Equation  be an infinite sequence of real-valued random variables and let N be a nonnegative integer-valued random variable.

Assume that:

    . (Xn)nWald's Equation  are all integrable (finite-mean) random variables,
    . E[Xn1{Nn}] = E[Xn] P(Nn) for every natural number n, and
    . the infinite series satisfies
        Wald's Equation 

Then the random sums

    Wald's Equation 

are integrable and

    Wald's Equation 

If, in addition,

    . (Xn)nWald's Equation  all have the same expectation, and
    . N has finite expectation,

then

    Wald's Equation 

Remark: Usually, the name Wald's equation refers to this last equality.

Discussion of assumptions

Clearly, assumption (1) is needed to formulate assumption (2) and Wald's equation. Assumption (2) controls the amount of dependence allowed between the sequence (Xn)nWald's Equation  and the number N of terms; see the counterexample below for the necessity. Note that assumption (2) is satisfied when N is a stopping time for a sequence of independent random variables (Xn)nWald's Equation .[citation needed] Assumption (3) is of more technical nature, implying absolute convergence and therefore allowing arbitrary rearrangement of an infinite series in the proof.

If assumption (5) is satisfied, then assumption (3) can be strengthened to the simpler condition

    . there exists a real constant C such that E[|Xn| 1{Nn}] ≤ C P(Nn) for all natural numbers n.

Indeed, using assumption (6),

    Wald's Equation 

and the last series equals the expectation of N [Proof], which is finite by assumption (5). Therefore, (5) and (6) imply assumption (3).

Assume in addition to (1) and (5) that

    . N is independent of the sequence (Xn)nWald's Equation  and
    . there exists a constant C such that E[|Xn|] ≤ C for all natural numbers n.

Then all the assumptions (1), (2), (5) and (6), hence also (3) are satisfied. In particular, the conditions (4) and (8) are satisfied if

    . the random variables (Xn)nWald's Equation  all have the same distribution.

Note that the random variables of the sequence (Xn)nWald's Equation  don't need to be independent.

The interesting point is to admit some dependence between the random number N of terms and the sequence (Xn)nWald's Equation . A standard version is to assume (1), (5), (8) and the existence of a filtration (Fn)nWald's Equation 0 such that

    . N is a stopping time with respect to the filtration, and
    . Xn and Fn–1 are independent for every nWald's Equation .

Then (10) implies that the event {Nn} = {Nn – 1}c is in Fn–1, hence by (11) independent of Xn. This implies (2), and together with (8) it implies (6).

For convenience (see the proof below using the optional stopping theorem) and to specify the relation of the sequence (Xn)nWald's Equation  and the filtration (Fn)nWald's Equation 0, the following additional assumption is often imposed:

    . the sequence (Xn)nWald's Equation  is adapted to the filtration (Fn)nWald's Equation , meaning the Xn is Fn-measurable for every nWald's Equation .

Note that (11) and (12) together imply that the random variables (Xn)nWald's Equation  are independent.

Application

An application is in actuarial science when considering the total claim amount follows a compound Poisson process

    Wald's Equation 

within a certain time period, say one year, arising from a random number N of individual insurance claims, whose sizes are described by the random variables (Xn)nWald's Equation . Under the above assumptions, Wald's equation can be used to calculate the expected total claim amount when information about the average claim number per year and the average claim size is available. Under stronger assumptions and with more information about the underlying distributions, Panjer's recursion can be used to calculate the distribution of SN.

Examples

Example with dependent terms

Let N be an integrable, Wald's Equation 0-valued random variable, which is independent of the integrable, real-valued random variable Z with E[Z] = 0. Define Xn = (–1)n Z for all nWald's Equation . Then assumptions (1), (5), (7), and (8) with C := E[|Z|] are satisfied, hence also (2) and (6), and Wald's equation applies. If the distribution of Z is not symmetric, then (9) does not hold. Note that, when Z is not almost surely equal to the zero random variable, then (11) and (12) cannot hold simultaneously for any filtration (Fn)nWald's Equation , because Z cannot be independent of itself as E[Z2] = (E[Z])2 = 0 is impossible.

Example where the number of terms depends on the sequence

Let (Xn)nWald's Equation  be a sequence of independent, symmetric, and {–1, +1}-valued random variables. For every nWald's Equation  let Fn be the σ-algebra generated by X1, . . . , Xn and define N = n when Xn is the first random variable taking the value +1. Note that P(N = n) = 1/2n, hence E[N] < ∞ by the ratio test. The assumptions (1), (5) and (9), hence (4) and (8) with C = 1, (10), (11), and (12) hold, hence also (2), and (6) and Wald's equation applies. However, (7) does not hold, because N is defined in terms of the sequence (Xn)nWald's Equation . Intuitively, one might expect to have E[SN] > 0 in this example, because the summation stops right after a one, thereby apparently creating a positive bias. However, Wald's equation shows that this intuition is misleading.

Counterexamples

A counterexample illustrating the necessity of assumption (2)

Consider a sequence (Xn)nWald's Equation  of i.i.d. (Independent and identically distributed random variables) random variables, taking each of the two values 0 and 1 with probability 1/2 (actually, only X1 is needed in the following). Define N = 1 – X1. Then SN is identically equal to zero, hence E[SN] = 0, but E[X1] = 1/2 and E[N] = 1/2 and therefore Wald's equation does not hold. Indeed, the assumptions (1), (3), (4) and (5) are satisfied, however, the equation in assumption (2) holds for all nWald's Equation  except for n = 1.[citation needed]

A counterexample illustrating the necessity of assumption (3)

Very similar to the second example above, let (Xn)nWald's Equation  be a sequence of independent, symmetric random variables, where Xn takes each of the values 2n and –2n with probability 1/2. Let N be the first nWald's Equation  such that Xn = 2n. Then, as above, N has finite expectation, hence assumption (5) holds. Since E[Xn] = 0 for all nWald's Equation , assumptions (1) and (4) hold. However, since SN = 1 almost surely, Wald's equation cannot hold.

Since N is a stopping time with respect to the filtration generated by (Xn)nWald's Equation , assumption (2) holds, see above. Therefore, only assumption (3) can fail, and indeed, since

    Wald's Equation 

and therefore P(Nn) = 1/2n–1 for every nWald's Equation , it follows that

    Wald's Equation 

A proof using the optional stopping theorem

Assume (1), (5), (8), (10), (11) and (12). Using assumption (1), define the sequence of random variables

    Wald's Equation 

Assumption (11) implies that the conditional expectation of Xn given Fn–1 equals E[Xn] almost surely for every nWald's Equation , hence (Mn)nWald's Equation 0 is a martingale with respect to the filtration (Fn)nWald's Equation 0 by assumption (12). Assumptions (5), (8) and (10) make sure that we can apply the optional stopping theorem, hence MN = SNTN is integrable and

    Wald's Equation 

    ()

Due to assumption (8),

    Wald's Equation 

and due to assumption (5) this upper bound is integrable. Hence we can add the expectation of TN to both sides of Equation (13) and obtain by linearity

    Wald's Equation 

Remark: Note that this proof does not cover the above example with dependent terms.

General proof

This proof uses only Lebesgue's monotone and dominated convergence theorems. We prove the statement as given above in three steps.

Step 1: Integrability of the random sum SN

We first show that the random sum SN is integrable. Define the partial sums

    Wald's Equation 

    ()

Since N takes its values in Wald's Equation 0 and since S0 = 0, it follows that

    Wald's Equation 

The Lebesgue monotone convergence theorem implies that

    Wald's Equation 

By the triangle inequality,

    Wald's Equation 

Using this upper estimate and changing the order of summation (which is permitted because all terms are non-negative), we obtain

    Wald's Equation 

    ()

where the second inequality follows using the monotone convergence theorem. By assumption (3), the infinite sequence on the right-hand side of (15) converges, hence SN is integrable.

Step 2: Integrability of the random sum TN

We now show that the random sum TN is integrable. Define the partial sums

    Wald's Equation 

    ()

of real numbers. Since N takes its values in Wald's Equation 0 and since T0 = 0, it follows that

    Wald's Equation 

As in step 1, the Lebesgue monotone convergence theorem implies that

    Wald's Equation 

By the triangle inequality,

    Wald's Equation 

Using this upper estimate and changing the order of summation (which is permitted because all terms are non-negative), we obtain

    Wald's Equation 

    ()

By assumption (2),

    Wald's Equation 

Substituting this into (17) yields

    Wald's Equation 

which is finite by assumption (3), hence TN is integrable.

Step 3: Proof of the identity

To prove Wald's equation, we essentially go through the same steps again without the absolute value, making use of the integrability of the random sums SN and TN in order to show that they have the same expectation.

Using the dominated convergence theorem with dominating random variable |SN| and the definition of the partial sum Si given in (14), it follows that

    Wald's Equation 

Due to the absolute convergence proved in (15) above using assumption (3), we may rearrange the summation and obtain that

    Wald's Equation 

where we used assumption (1) and the dominated convergence theorem with dominating random variable |Xn| for the second equality. Due to assumption (2) and the σ-additivity of the probability measure,

    Wald's Equation 

Substituting this result into the previous equation, rearranging the summation (which is permitted due to absolute convergence, see (15) above), using linearity of expectation and the definition of the partial sum Ti of expectations given in (16),

    Wald's Equation 

By using dominated convergence again with dominating random variable |TN|,

    Wald's Equation 

If assumptions (4) and (5) are satisfied, then by linearity of expectation,

    Wald's Equation 

This completes the proof.

Further generalizations

  • Wald's equation can be transferred to Rd-valued random variables (Xn)nWald's Equation  by applying the one-dimensional version to every component.
  • If (Xn)nWald's Equation  are Bochner-integrable random variables taking values in a Banach space, then the general proof above can be adjusted accordingly.

See also

Notes

References

Tags:

Wald's Equation Basic versionWald's Equation ExampleWald's Equation General versionWald's Equation Discussion of assumptionsWald's Equation ApplicationWald's Equation ExamplesWald's Equation CounterexamplesWald's Equation A proof using the optional stopping theoremWald's Equation General proofWald's Equation Further generalizationsWald's EquationExpected valueIdentity (mathematics)Independence (probability theory)Independent and identically distributed random variablesProbability theory

🔥 Trending searches on Wiki English:

Mariah CareyJohnny CashSri LankaAaron HernandezMukesh AmbaniCable (character)FlipkartGeorge W. BushDev PatelJordan CameronRomelu LukakuArgentina national football teamSeth MacFarlaneIranLimoneneAlbert EinsteinCillian MurphyIslamic StateCrucifixion of JesusCristiano RonaldoMaster X Master1xBetCorfuGoogle TranslateElena RybakinaWes MooreXNXXMurder MubarakI-35W Mississippi River bridgeDeclan RiceMichael Jackson3 Body Problem (TV series)Ted BundyContinuous truss bridgeRichard SerraSuge KnightAmerican Civil WarCheryl HinesTokugawa IeyasuTimothée ChalametEmily BluntLove Lies Bleeding (2024 film)2024 Indian general election in KarnatakaGeneral averageMel GibsonRocco SiffrediWorld War IIGrey's AnatomyPreity ZintaList of presidents of the United States2024 Indian general election in Tamil NaduCarlos Sainz Jr.Daniela MelchiorCorey HaimWillie NelsonList of Hindi films of 2024Red yeast riceJosh PeckDebbie ReynoldsLisa Marie Presley2024 Indian Premier LeaguePeaky Blinders (TV series)Gujarat TitansTom HollandFacebookGame Changer (film)Russo-Ukrainian WarTom CruiseThe Walking Dead (TV series)VietnamIsraelRegina KingNetEaseJulia RobertsRashida JonesTiger Woods🡆 More