A stochastic process is a collection of random variables defined on a probability space and assuming values in with . The index set is usually the half-like , but can be also an interval or a subset of .
While a stochastic process has a crystal clear, mathematical definition. A time series is a less precise notion, and people use time series to refer to two related but different objects. A time series can be seen as a stochastic process indexed by integers or some regular, incremental unit of time that can in a sense by mapped to integers (eg. minutely, hourly, daily, monthly data). A time series can be understood as a collection of time-value–data-point pairs. while a stochastic process is the mathematical description of a distribution of time series. Some time series are a realization of stochastic processes or one can use a stochastic process as a model to generate a time series.
Definition 18.1 ()
Let be a probability space and let be an index set. If for every , one have that is a sub--algebra of and for every holds , then the family of sub--algebras denoted with is called a filtration.
For example, considering a stochastic process and consider the sequence of sub--algebras generated by , then is a -algebra and is a filtration, also called natural filtration, with respect to . Formally, is the smallest -algebra containing all events observable up to the index , i.e.
18.1 Stationarity
Definition 18.2 ()
A stochastic process is said to be strongly stationary if and only if for all set of index and for every meaning that the joint distribution of an arbitrary number of random variables do not change shifting the process, upward or downward, by a step .
Definition 18.3 ()
A stochastic process is said to be weakly stationary (or covariance stationary) if and only if
and for every ;
for every ;
and for every and .
Hence, the covariance of a weakly stationary process does not depend on time , but only on the temporal lag between two observations.
Strong does not imply weakly and viceversa
In general if a process is strong stationary (Definition 18.2) does not implies that it is also weakly stationary. For example, an independent and identically distributed Cauchy process is strongly stationary, but since its expectation and variance are not finite the process is not weakly stationary.
18.2 Notable processes
Definition 18.4 ()
A time series, where each is independent from the others and all has the same distribution for all is called independent and identically distributed process (IID). Such kind of process, usually denoted as , is strongly stationary (Definition 18.2). Moreover, if the mean and variance are finite, the covariance is zero and the process is also weakly stationary (Definition 18.3), i.e.
Definition 18.5 ()
A stochastic process , commonly denoted as is called White Noise if satisfies the following properties:
The expectation is equal to zero, i.e. for all .
The variance is finite and constant for all , i.e. .
The process is uncorrelated over time for all , i.e. .
A White Noise process is weakly stationary (Definition 18.3). In fact, the autocovariance function of the process depends on the lag, but not on time, i.e. it is equal to the variance for and is zero otherwise. This process is more general than an IID process (Definition 18.4), since it does not requires the stochastic independence of the time series for all .
18.2.1 Martingales
Definition 18.6 () Let’s consider a probability space and stochastic process . Then, given a filtration , namely , the stochastic process is a martingale with respect to the filtration if
is adapted to in the sense that in included in the information contained in , i.e. is -measurable.
For any
Definition 18.7 ()
Let’s consider a probability space and stochastic process . Then, given a filtration , namely , the stochastic process is a martingale difference sequence (MDS) with respect to the filtration if is -measurable and for any This implies that is a mean-zero process uncorrelated with any information contained . The definition can be extended to a case where the filtration includes also other processes . In this case, is said to be an MDS conditionally to if the same condition in Equation 18.3 holds.
Super and sub-martingales
A stochastic process is said to be a sub-martingale if instead of Equation 18.2 we have that for any , On the other hand it is said to be a super-martingale if From the above definitions it follows that to be a martingale (or an MDS) a stochastic process must be at the same time a super and sub-martingale.
Simulate Martingales
set.seed(1)# Number of observationsN <-100# Number of simulationsj_bar <-20# Predictable processA_t <-rep(0.2, N)X_sub <- X_mar <- X_sup <-list()for(j in1:j_bar){# Martingale M_t <-rnorm(N, 0, 1)# Sub-martingale X_sub[[j]] <- dplyr::tibble(j = j, t =1:N, X =cumsum(A_t + M_t))# Martingale X_mar[[j]] <- dplyr::tibble(j = j, t =1:N, X =cumsum(M_t))# Super-martingale X_sup[[j]] <- dplyr::tibble(j = j, t =1:N, X =cumsum(-A_t + M_t))}
Figure 18.1: Simulation of a sub-martingale, martingale and super-martingale with expected value (red).
The concept of martingales is connected to the concept of predictability of a stochastic process.
Definition 18.8 () Let’s consider a probability space and stochastic process . Then, let’s define a sequence of sub- fields of , namely . Then, the stochastic process is predictable if
;
For any we have that .
Then, we call the process predictable and increasing if .
Theorem 18.1 () Any sub-martingale , can be written in a unique way as the sum of a martingale and an increasing process , i.e.
18.3 Lag operator
The lag operator is a function that allows to translate a time series in time. In general, the lag operator associate at it’s lagged value , i.e. More formally, is the operator that takes one whole time series and produces another; the second time series is the same as the first, but moved backwards or forward one point in time. From the definition, we list some properties related to the Lag operator, i.e.
Backward.
Forward.
.
18.3.1 Polynomial of Lag operator
Given a time series , it is possible to define polynomials of the Lag operator, i.e. where in general
For the polynomial holds the factorization where are the complex solutions of the characteristic equation, i.e. Hence the factorization holds true if and only if: In other words, the modulus of the solutions must outside the unit circle, otherwise the geometric series is not convergent and the factorization do not holds true anymore. The factorization of the lag polynomial allows us to define its inverse, i.e. In fact, the inverse of the -th term can be expressed with a Taylor expansion as infinite sum if and only if , i.e. that is equivalent to for all since .
AR(1) and geometric series
For example, let’s consider an Autoregressive process of order 1, i.e. In fact, Considering such polynomial, its inverse polynomial , defined such that , is defined as geometric series, i.e. that converges if and only if . Moreover, if it is possible to prove that is indeed the inverse polynomial of , in fact: Therefore, the process can be equivalently expressed as: The factorization of any polynomial of the form of is connected to the convergence of the following geometric series, i.e.
Convergent series (I)
# *************************************************# Inputs # *************************************************a <-c(0.7, -0.7)min_i <-0max_i <-20i <-seq(min_i, max_i -1, 1)y_breaks <-seq(min_i, max_i, 2)x_labels <-quantile(i, 0.85)# *************************************************# Convergent series 0 < a < 1series1 <-cumsum(a[1]^i)limit1 <-1/(1- a[1])# Convergent series -1 < a < 0series2 <-cumsum(a[2]^i)limit2 <-1/(1- a[2])
(a) .
(b) .
Figure 18.2: Convergent series for AR(1) parameter (I).
Another important series that is convergent only if and only if , i.e. Due to the square, in this case we do not distinguish between and since they lead to the same result.
Convergent series (II)
# Convergent series |a| < 1 series1 <-cumsum(a[1]^(i*2))limit1 <-1/(1- a[1]^2)
Figure 18.3: Convergent series for AR(1) parameter (II).