7  Characteristic functions

Definition 7.1 (Characteristic function)
Consider an n-dimensional vector X, then the characteristic function is defined tRn as: ΦX(t)=E{eitTX}wheretTX=(t1,t2,,tn)(X1X2Xn)

The characteristic function always exists when treated as a function of a real-valued argument, unlike the moment-generating function. The characteristic function uniquely determines the probability distribution of the correspondent random vector X. More precisely, saying that two random variables has the same distribution is equivalent to say that their characteristic functions are equal. It follows that we can always work under characteristic functions to prove that two distribution of some random vectors are equal or that a distribution converges to another distribution. Formally X and Y ave same distribution, i.e.  XYΦX(t)=ΦY(t)tRn

Here, we list some properties considering the random variable case, i.e. n=1, with tR.

  1. Independence: X and Y are independent iff: XYΦX+Y(t)=ΦX(t)ΦY(t)tR

  2. Existence of the j-th moment: If the j-th moment of the random variable is finite then the characteristic function is j-times differentiable and continuous in 0, i.e. ΦXC(j). Formally, E{|X|r}<ΦXC(r)andΦX(r)(t)=E{(iX)reitX}r=1,2,,j Note that, if j is even, it became an if: E{|X|r}<ΦXC(r),ΦX(r)(t)=E{(iX)reitX}r=2,4,,j

  3. Inversion theorem: The characteristic function uniquely determines the probability distribution of the correspondent random vector X: F(b)F(a)=12πilimccceitaeitbtΦ(t)dta<b Then, the density function is obtained as: f(x)=12πeitXΦX(t)dt

  4. Convergence in distribution: XnndXΦX(t)=limnΦXn(t)tR

  5. Scaling and centering: Given Y=a+bX, the effect of scaling and centering on the characteristic function is such that: ΦY(t)=E{eitY}=E{eit(a+bX)}=eitbE{ei(ta)X}=eitbΦX(at)a,b,tR

  6. Weak Law of Large Numbers: consider a sequence of IID random variables {Xn}n1, such that exists the first derivative of the characteristic function in zero, namely ϕX(0) and the first moment of X1 is finite, i.e. E{|X1|}=μ<, then the sample mean converges in probability to a degenerate random variable αR. X¯n=1ni=1nXipnα for some αR.

WLLN and characteristic function

Proof. To prove the above statement, let’s compute the characteristic function of X¯n: ϕX¯n(t)=E{exp(itni=1nXn)}=IID[ϕX1(tn)]n Let’s now apply the Taylor series of a function f(x) around the point a () to expand till the first order term the function ϕX1(tn) around zero (a=0 and x=tn), i.e. ϕX¯n(tn)=(ϕX1(0)+tnϕX1(0)+o(tn))n==(1+tϕX1(0)+no(tn)n)nnexp{tϕX1(0)} The convergence going to the limit as n follows from the fact that in general if anna, then the following limit holds: (1+ann)nnea Therefore, since ϕX1(0)=iα for some αR, it follows that: limnϕX¯n(t)=etϕX1(0)=eitαtR Hence, since eitα is the characteristic function of a degenerate random variable, namely a random variable that is constant almost surely, it is possible to conclude that the sample mean converges in distribution to a degenerate random variable α. Moreover, in this specific case in which the limit is a degenerate random variable it can be proved that having convergence in distribution implies also convergence in probability, something that in general is not true.

7.1 Moment generating function

Definition 7.2 (Moment Generating Function)
Consider an uni dimensional random variable X, then the moment generating function is defined as: ψX(t)=E{etX}tR{0}

Proposition 7.1 (Moment generating function and sequence of moments)
Consider a random variable X, such that it’s moment generating function exists and it’s finite around zero, i.e. ψX(t)=E{etX}<ϵ>0,t(ϵ,ϵ) Then this implies that the sequence of moments are finite E{|X|n}< for all n and the sequence of moments uniquely determine the distribution of X. According to this result, if we consider another random variable Y such that E{|X|n}=E{|Y|n} for all n, then the distribution of X and Y is the same, i.e. XY.