5  Expectation

The expectation represents a central value of a random variable and has a measure theory counterpart as a Lebesgue-Stieltjes integral of X with respect to a (probability) measure P. This kind of integration is defined in steps. First it is shown the integration of simple functions and then extended to more general random variables. In general, let’s consider a probability space (Ω,B,P) and a random variable X such that X:(Ω,B)(R¯,B(R¯)) where R¯=[,]. Then, the expectation of X is denoted as: E{X}=ΩXdP=ΩX(ω)P(dω) as the Lebesgue-Stieltjes integral of X with respect to the (probability) measure P.

5.1 Simple functions

In general a random variable X(ω) is simple if it has a finite range. Let’s consider a probability space (Ω,B,P) and consider a B/B(R)-measurable simple function X:ΩR, i.e.  (5.1)X(ω)=i=1nai1Ai(ω), where aiR and AiB are a disjoint partition of the sample space, i.e. i=1nAi=Ω. Let’s denote the set of all simple functions on Ω as E. In this settings, E is a vector space. This implies that he following two properties holds.

  1. Constant: given a simple function XE, then αXE. In fact: (5.2)αX=i=1nαai1Ai=i=1nai1AiE where ai=αai.

  2. Linearity: given two simple function X,YE, then X+YE. In fact: (5.3)X+Y=i=1nai1Ai+j=1mbj1Bj==i=1nj=1m(ai+bj)1Ai1Bj==i=1nj=1m(ai+bj)1AiBj where the sequence of sets {AiBj1inand1jm} form a disjoint partition of Ω.

  3. Product: given two simple function X,YE, then XYE. In fact: (5.4)XY=i=1nai1Aij=1mbj1Bj==i=1nj=1m(aibj)1Ai1Bj==i=1nj=1m(aibj)1AiBj

5.1.1 Measurability

Simple functions are the building blocks in the definition of the expectation in terms of Lebesgue-Stieltjes integral. In fact a known theorem called Measurability theorem shows that any measurable function can be approximated by a sequence of simple functions.

Theorem 5.1 Suppose that X(ω)0 for all ωΩ. Then, X is B/B(R) measurable if and only if there exists simple functions XnE and 0XnXX=limnXn

5.2 Expectation of Simple Functions

The expectation of a simple function X is defined as: E{X}=i=1naiP(Ai) where |ai|<.

5.2.1 Properties

  1. Non-negativity: If X0 and XE then E{X}0

Proof. By definition of simple functions

  1. Linearity: the expectation of simple function is linear, i.e.  E{αX+βY}=αE{X}+βE{Y}

Proof. Let’s consider two simple functions, i.e.  X(ω)=i=1nai1Ai(ω)andY(ω)=j=1mbj1Bj(ω), and let’s fix α,βR. Then, by the second property of the vector space E () it is possible to write: αX+βY=i=1nj=1m(αai+βbj)1AiBj Then, taking the expectation on both sides:
E{αX+βY}=i=1nj=1m(αai+βbj)P(AiBj)==i=1nαaij=1mP(AiBj)+j=1mβbji=1nP(AiBj) Fixing i, the sequence AiBj for j=1,,n is composed by disjoint events since by definition Bj are disjoint. Hence, applying σ-additivity it is possible to write: j=1mP(AiBj)=P(j=1mAiBj)==P(Ai(j=1mBj))==P(AiΩ)=P(Ai) Therefore, the expectation simplifies in: E{αX+βY}=i=1nαaiP(Ai)+j=1mβbjP(Bj)==αE{X}+βE{Y}

5.3 Review of inequalities

5.3.1 Modulus inequality

Definition 5.1 (Modulus Inequality)
Let’s consider a random variable XL1, where L1 stands for the set of integrable random variables, i.e.  L1={X:ΩR:X is a r.v. ,E{|X|}<} Then, the modulus inequality states that: |E{X}|E{|X|}

5.3.2 Markov inequality

Definition 5.2 (Markov Inequality)
Let’s consider a random variable XL1 and fix a λ>0, then by the Markov inequality: P(|X|λ)1λE{|X|}

5.3.3 Chebychev inequality

Definition 5.3 (Chebychev Inequality)
Consider a random variable X with first and second moment finite, i.e.  E{|X|}<,V{|X|}< then by the Chebychev inequality: (5.5)P(Xλ)1λ2E{|X|2}

5.3.4 Holder inequality

Definition 5.4 (Holder Inequality)
Let’s consider two numbers p and q such that p>1,q>1,1p+1q=1 and let’s consider two random variables X and Y such that: E{|X|p}<,E{|Y|q}< Then, (5.6)|E{XY}|E{|XY|}(E{|X|p})1p(E{|Y|q})1q In terms of norms: ||XY||1||X||p||Y||q

5.3.5 Schwartz inequality

Definition 5.5 (Schwartz Inequality)
Consider two random variables X,YL2, i.e. with first and second moment finite, i.e.  E{|X|}<,E{X2}< Then (5.7)|E{XY}|E{|XY|}E{X2}E{Y2} In terms of norms: ||XY||1||X||2||Y||2 Note that this is a special case of Holder inequality () with p=q=2.

5.3.6 Minkowski inequality

Definition 5.6 (Minkowski Inequality)
For 1p< let’s consider two random variables X,YLp, then X+YLp and (5.8)||X+Y||p||X||p+||Y||p

Note that the triangular inequality is a special case of Minkowski inequality with p=1, i.e.  (5.9)|X+Y||X|+|Y|

5.3.7 Jensen inequality

Definition 5.7 (Jensen Inequality)
Let’s consider a convex function u:RR. Suppose that E{X}< and E{|u(X)|}<, then (5.10)E{u(X)}u(E{X}) if u is concave the results revert, i.e.  (5.11)E{u(X)}u(E{X})