Reference: Pietro Rigo (2023)
Theorem 6.1 ()
Consider a measure space and two measures , such that is -finite (Definition 30.4) and (Definition 30.1). Then there exists a measurable function such that:
Definition 6.1 ()
Given a probability space . Let’s consider a sub -field of , i.e. and a random variable with finite expectation . Then, the conditional expectation of given is any random variable such that:
- has finite expectation, i.e. .
- is -measurable.
- , , namely if and are restricted to , then their expectation coincides.
A -field can be used to describe our state of information. It means that, we already know if the event has occurred or not. Therefore, when we insert in the events that we know were already occurred, we are saying that the random variable is -measurable, i.e. the value of is not stochastic once we know the information contained in . In this context, one can see the random variable as the prediction of , given the information contained in the sub -field .
Definition 6.2 ()
Consider any -measurable random variable. Then can be interpreted as a predictor of another random variable under the information contained in the -field . When we substitute with its prediction , we make an error given by the difference . In the special case in which and using as error function the mean squared error, i.e. then it is possible to prove that the conditional expectation represent the best predictor of in the sense that it minimized the mean squared error, i.e. Hence, is the best predictor that minimize the mean squared error over the class composed by -measurable functions with finite second moment, i.e. where .
Properties of conditional expectation
Here we state some useful properties of conditional expectation:
- Linearity: The conditional expectation is linear for all all constants , i.e.
- Positive: implies that .
- Measurability: If is -measurable, then . In general, if is -measurable then , i.e. is not stochastic.
- Constant: The conditional expectation of a constant is a constant, i.e.
- Independence: If is independent from the -field , then .
- Chain rule: If one consider two sub -fields of such that , then we can write: Remember that, when using the chain rule it is mandatory to take the conditional expectation before with respect to the greatest -field, i.e. the one that contains more information (in this case ), and then with respect to the smallest one (in this case ).
Conditional probability
Proposition 6.1 ()
Given a probability space , consider as a sub -field of , i.e. . Then the general definition of the conditional probability of an event , given , is Instead, the elementary definition do not consider the conditioning with respect to a -field, but instead with respect to a single event . In practice, take an event such that , then the conditional probability of given is defined as:
Proof. The elementary (Equation 6.3) and the general (Equation 6.2) definitions are equivalent, in fact consider a sub -field which provides only the information concerning whenever is in or not. A -field of this kind will have the form . Then, consider a -measurable function, , such that: It remains to find and in the following expression: Note that, the joint probability of and can be obtained as: Hence, we obtain: Equivalently for it is possible to prove that: Finally, thanks to this result it is possible to express the conditional probability in the general definition (Equation 6.2) as a linear combination of conditional probabilities defined accordingly to the elementary definition (Equation 6.3), i.e.
Exercise 6.1 Let’s continue from the Exercise 3.1 and let’s say that we observe , then we ask ourselves, what is the probability that in the next extraction ?
Solution 6.1. The chances that with 52 cards we obtain is approximately (see Solution 3.2), while is approximately .
Hence, if we have extracted a card that originates an , then in the deck remain only 19 possible cards that gives while the number of total cards reduces to 51. Thus, the conditional probability on another 1 decreases On the other hand, the conditional probability of a 0 increases
Proposition 6.2 ()
If two events and are independent and , then
Proof. To prove the result in Proposition 6.2 we consider both sides of the expression.
(LHS RHS) Let’s assume that and are independent, i.e. . Then, by definition (Definition 4.1) under independence holds the decomposition in Equation 4.1, i.e.
(RHS LHS) Let’s assume that the following holds, i.e. and . Then, by definition of conditional probability (Equation 6.3) the joint probability is defined as: that is exactly the definition of independence (Definition 4.1).
Theorem 6.2 ()
Let’s consider a partition of disjoint events with each and such that . Then, given any event with probability greater than zero, . Then, for any the conditional probability of the event given is defined as:
Example 6.1 Let’s consider two random variables and taking values in . The marginal probabilities and . Let’s consider the matrix of joint events and probabilities, i.e. Then, by definition the conditional probabilities are defined as: and Considering instead: and Then, it is possible to express the marginal probability of as: And similarly for
Exercise 6.2 ()
You are on a game show where there are three closed doors: behind one is a car (the prize) and behind the other two are goats.
The rules are simple:
- You choose one door (say Door 1).
- The host, who knows where the car is, opens one of the other doors, always revealing a goat.
- You are then offered the chance to stay with your original choice or switch to the other unopened door.
Question: Is in your interests to switch door? (See 21 Blackjack)
Solution 6.2. Before any door is opened, the probability that the car is behind each door is Suppose you picked Door 1. The conductor opens (say) Door 3, revealing a goat. Now the conditional probabilities are:
If the car is behind Door 1: Monty could open either Door 2 or Door 3 with equal probability.
If the car is behind Door 2: Monty is forced to open Door 3.
If the car is behind Door 3: Monty is forced to open Door 2 (so this case is impossible if Monty opens Door 3).
Apply Bayes’ Rule: On the other hand, the other door has probabvility of winning of After Monty opens a goat door, the probability the car is behind your original choice is still 33, while the probability it is behind the other unopened door is 66, almost double . Therefore, switching doubles the chances of winning.
Conditional variance
Proposition 6.3 ()
Let’s consider two random variable and with finite second moment. Then, the total variance can be expressed as:
Proof. By definition, the variance of a random variable reads: Applying the chain rule (Equation 6.1) one can write Then, add and subtract gives Group the first and fourth terms and the second and third to obtain that simplifies as
Pietro Rigo. 2023. “Appunti Di Probabilità (PhD Courses).”