A measurable space is composed by a sample space and a -field of subsets of , namely .
3.1 Maps and inverse maps
Let’s be very general and consider two measurable spaces , and a map (function) that associate to every an outcome , i.e. Then, determine a function called inverse map, i.e. where denotes the power set, i.e. the set of all subsets and the largest -field available. Then, is defined such that for every set in
Exercise 3.1 Let’s consider a deck of poker cards with 52 cards in total. We have 4 groups of 13 distinct cards, where the Jack (J) is 11, the Queen (Q) is 12, the King (K) is 13 and Ace (A) is 14. Then, let’s consider function of the form In this case the sample space is composed by 14 unique elements (52 elements in total, i.e. all the cards), while represents the possible outcomes. Let’s say that one observe a value , define according to Equation 3.1.
Solution 3.1. Let’s say that one observe a value , then the inverse map identifies the set of such that , i.e.
The inverse map has many properties. Among them, it preserves complementation, union and intersection.
.
.
.
.
for all .
Properties of inverse maps
Let’s consider two sets and both in . Then, by definition: Similarly for the intersection, i.e.
Proposition 3.1 If is a -field of subsets of , then is a -field of subsets of . Moreover, if is a class of subsets of , then that is. the inverse image of the -field generated by the class is the same as the -field generated in by the inverse image . In practice, the counter image and the generators commute. Usually can be difficult to know all about the -field , however if we know a class of subset that generate it, namely , we are able to recreate the -field. (References: propositions 3.1.1, 3.1.2. A prob path).
3.1.1 Measurable maps
Definition 3.1 ()
Let’s consider the function , then is -measurable, namely , iff:
Note that the measurability concept is very important since only if is measurable it is possible to make probability statements about . In fact, the probability is nothing more than a type of measure, only if for all it is possible to assign a measure (probability) to the events that are contained in .
Definition 3.2 ()
Consider a map and the class that generates the -field , i.e. . Then is -measurable iff:
3.2 Random variables
Definition 3.3 ()
Let’s consider a probability space , then a random variable is a map (function) where . Therefore it takes values on the real line, i.e. and for every set in the Borel -field generated by the real line , the counter image of , namely is in the -field generated by . Formally, More precisely, when is a random variable the test for measurability (Definition 3.2), became: In practice, that depends on the real number while the event is exactly the counter image of .
3.2.1-field generated by a map
Let be a measurable map, then the -field generated by is defined as: When is a random variable, then and the -field generated by
Definition 3.4 Let be a random variable with set of possible outcomes . Then, is called
discrete random variable if is either a finite set or a countably infinite set.
continuous random variable if is either a an uncountable infinite set.
Definition 3.5 A random vector is a map It’s counter image is defined as: Note that, for every , the random vector has -components, i.e.
Proposition 3.2 The map is a random vector, if and only if each component is a random variable.
Proof. Let’s prove that if are random variables, then is a random vector. Since we are assuming that each -th component is a random variable we have that for all Therefore, to prove that is a random vector we have to prove that the cartesian product of all the counter images, i.e. is in . That is equivalent to the intersection Since, each the counter image of each and is a -field, hence closed under countable intersection, also the intersection of all the counter images will be in .
3.3 Induced distribution function
Consider a probability space and a measurable map , then the composition is again a map. In this way at each element is attached a probability measure. In fact, the composition is a map such that In general, the probability of a subset is denoted equivalently as:
Exercise 3.2 Let’s continue from the Exercise 3.1 and compute the probability of . Let’s consider one random extraction from the 52 cards, then for each distinct number we have 4 copies. Compute and .
Solution 3.2. The probability is computed as: Let’s consider the probability of observing either or , then Finally, by property of the probability measure .
3.3.1 Distribution function on
When is a random variable the composition is a probability measure induced on by the composition: for all in . Hence, the distribution function of a random variable is a function and represents a probability measure on the real line , i.e. or for short .
Warning
In general, the probability distribution is a function , however when is a random variable this means that .
(a) Distribution functions.
(b) Quantile functions.
Figure 3.1: Different distribution functions in .
Proposition 3.3 ()
If a random variable has a continuous and differentiable distribution function, then its probability density function (pdf) is defined as the first derivative of the distribution function with respect to , i.e. Instead, for a discrete random variable we call it probability mass function (pmf), i.e. Considering a generic domain for the random variable , then the function satisfies two fundamental properties, i.e.
Positivity: for all .
Normalization: or .
Warning
In general, any function that satisfies properties 1. and 2. in Proposition 3.3 is the density function of some (unknown) random variable.
3.3.2 Survival function
For any random variable with distribution , the survival distribution is defined as:
Definition 3.6 ()
A distribution function is said to be: