SUPPORT THE WORK

GetWiki

stationary process

ARTICLE SUBJECTS
aesthetics  →
being  →
complexity  →
database  →
enterprise  →
ethics  →
fiction  →
history  →
internet  →
knowledge  →
language  →
licensing  →
linux  →
logic  →
method  →
news  →
perception  →
philosophy  →
policy  →
purpose  →
religion  →
science  →
sociology  →
software  →
truth  →
unix  →
wiki  →
ARTICLE TYPES
essay  →
feed  →
help  →
system  →
wiki  →
ARTICLE ORIGINS
critical  →
discussion  →
forked  →
imported  →
original  →
stationary process
[ temporary import ]
please note:
- the content below is remote from Wikipedia
- it has been imported raw for GetWiki
{{Short description|Type of stochastic process}}In mathematics and statistics, a stationary process (or a strict/strictly stationary process or strong/strongly stationary process) is a stochastic process whose unconditional joint probability distribution does not change when shifted in time.BOOK, Markov Chains: From Theory to Implementation and Experimentation, Gagniuc, Paul A., John Wiley & Sons, 2017, 978-1-119-38755-8, USA, NJ, 1–256, Consequently, parameters such as mean and variance also do not change over time.In other words, a line drawn through the middle of a stationary process — i.e. the trend line — is flat. It may have 'seasonal' cycles around this trend line, but overall it does not trend up nor down.Since stationarity is an assumption underlying many statistical procedures used in time series analysis, non-stationary data are often transformed to become stationary. The most common cause of violation of stationarity is a trend in the mean, which can be due either to the presence of a unit root or of a deterministic trend. In the former case of a unit root, stochastic shocks have permanent effects, and the process is not mean-reverting. In the latter case of a deterministic trend, the process is called a trend-stationary process, and stochastic shocks have only transitory effects after which the variable tends toward a deterministically evolving (non-constant) mean.A trend stationary process is not strictly stationary, but can easily be transformed into a stationary process by removing the underlying trend, which is solely a function of time. Similarly, processes with one or more unit roots can be made stationary through differencing. An important type of non-stationary process that does not include a trend-like behavior is a cyclostationary process, which is a stochastic process that varies cyclically with time.For many applications strict-sense stationarity is too restrictive. Other forms of stationarity such as wide-sense stationarity or N-th-order stationarity are then employed. The definitions for different kinds of stationarity are not consistent among different authors (see Other terminology).

Strict-sense stationarity

Definition

Formally, let left{X_tright} be a stochastic process and let F_{X}(x_{t_1 + tau}, ldots, x_{t_n + tau}) represent the cumulative distribution function of the unconditional (i.e., with no reference to any particular starting value) joint distribution of left{X_tright} at times t_1 + tau, ldots, t_n + tau. Then, left{X_tright} is said to be strictly stationary, strongly stationary or strict-sense stationary ifBOOK, Park,Kun Il, Fundamentals of Probability and Stochastic Processes with Applications to Communications, Springer, 2018, 978-3-319-68074-3, {{rp|p. 155}}{{Equation box 1|indent =|title=
0}>{{EquationRef|Eq.1}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}Since tau does not affect F_X(cdot), F_{X} is independent of time.

Examples

File:Stationarycomparison.png|thumb|right|390px|Two simulated time series processes, one stationary and the other non-stationary, are shown above. The augmented Dickey–Fuller (ADF) test statistic is reported for each process; non-stationarity cannot be rejected for the second process at a 5% significance levelsignificance levelWhite noise is the simplest example of a stationary process.An example of a discrete-time stationary process where the sample space is also discrete (so that the random variable may take one of N possible values) is a Bernoulli scheme. Other examples of a discrete-time stationary process with continuous sample space include some autoregressive and moving average processes which are both subsets of the autoregressive moving average model. Models with a non-trivial autoregressive component may be either stationary or non-stationary, depending on the parameter values, and important non-stationary special cases are where unit roots exist in the model.

Example 1

Let Y be any scalar random variable, and define a time-series left{X_tright}, by
X_t=Y qquad text{ for all } t.
Then left{X_tright} is a stationary time series, for which realisations consist of a series of constant values, with a different constant value for each realisation. A law of large numbers does not apply on this case, as the limiting value of an average from a single realisation takes the random value determined by Y, rather than taking the expected value of Y.The time average of X_t does not converge since the process is not ergodic.

Example 2

As a further example of a stationary process for which any single realisation has an apparently noise-free structure, let Y have a uniform distribution on [0,2pi] and define the time series left{X_tright} by
X_t=cos (t+Y) quad text{ for } t in mathbb{R}.
Then left{X_tright} is strictly stationary since ( (t+ Y) modulo 2 pi ) follows the same uniform distribution as Y for any t .

Example 3

Keep in mind that a weakly white noise is not necessarily strictly stationary. Let omega be a random variable uniformly distributed in the interval (0, 2pi) and define the time series left{z_tright}z_t=cos(tomega) quad (t=1,2,...) Then
begin{align}mathbb{E}(z_t) &= frac{1}{2pi} int_0^{2pi} cos(tomega) ,domega = 0,operatorname{Var}(z_t) &= frac{1}{2pi} int_0^{2pi} cos^2(tomega) ,domega = 1/2,operatorname{Cov}(z_t , z_j) &= frac{1}{2pi} int_0^{2pi} cos(tomega)cos(jomega) ,domega = 0 quad forall tneq j.end{align}So {z_t} is a white noise in the weak sense (the mean and cross-covariances are zero, and the variances are all the same), however it is not strictly stationary.{{clear}}

Nth-order stationarity

In {{EquationNote|Eq.1}}, the distribution of n samples of the stochastic process must be equal to the distribution of the samples shifted in time for all n. N-th-order stationarity is a weaker form of stationarity where this is only requested for all n up to a certain order N. A random process left{X_tright} is said to be N-th-order stationary if:{{rp|p. 152}}{{Equation box 1|indent =|title=
{{EquationRef|Eq.2}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}

Weak or wide-sense stationarity

Definition

A weaker form of stationarity commonly employed in signal processing is known as weak-sense stationarity, wide-sense stationarity (WSS), or covariance stationarity. WSS random processes only require that 1st moment (i.e. the mean) and autocovariance do not vary with respect to time and that the 2nd moment is finite for all times. Any strictly stationary process which has a finite mean and covariance is also WSS.BOOK, Ionut Florescu, Probability and Stochastic Processes, 7 November 2014, John Wiley & Sons, 978-1-118-59320-2, {{rp|p. 299}}So, a continuous time random process left{X_tright} which is WSS has the following restrictions on its mean function m_X(t) triangleq operatorname E[X_t] and autocovariance function K_{XX}(t_1, t_2) triangleq operatorname E[(X_{t_1}-m_X(t_1))(X_{t_2}-m_X(t_2))]:{{Equation box 1|indent =|title=
< infty & & text{for all } t in mathbb{R}end{align}
|begin{align}& m_X(t) = m_X(t + tau) & & text{for all } tau,t in mathbb{R} & K_{XX}(t_1, t_2) = K_{XX}(t_1 - t_2, 0) & & text{for all } t_1,t_2 in mathbb{R} & operatorname E[|X_t|^2] Eq.3}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}The first property implies that the mean function m_X(t) must be constant. The second property implies that the autocovariance function depends only on the difference between t_1 and t_2 and only needs to be indexed by one variable rather than two variables.{{rp|p. 159}} Thus, instead of writing,
,!K_{XX}(t_1 - t_2, 0),
the notation is often abbreviated by the substitution tau = t_1 - t_2:
K_{XX}(tau) triangleq K_{XX}(t_1 - t_2, 0)
This also implies that the autocorrelation depends only on tau = t_1 - t_2, that is
,! R_X(t_1,t_2) = R_X(t_1-t_2,0) triangleq R_X(tau).
The third property says that the second moments must be finite for any time t.

Motivation

The main advantage of wide-sense stationarity is that it places the time-series in the context of Hilbert spaces. Let H be the Hilbert space generated by {x(t)} (that is, the closure of the set of all linear combinations of these random variables in the Hilbert space of all square-integrable random variables on the given probability space). By the positive definiteness of the autocovariance function, it follows from Bochner's theorem that there exists a positive measure mu on the real line such that H is isomorphic to the Hilbert subspace of L2(μ) generated by {e−2{{pi}}iξ⋅t}. This then gives the following Fourier-type decomposition for a continuous time stationary stochastic process: there exists a stochastic process omega_xi with orthogonal increments such that, for all t
X_t = int e^{- 2 pi i lambda cdot t} , d omega_lambda,
where the integral on the right-hand side is interpreted in a suitable (Riemann) sense. The same result holds for a discrete-time stationary process, with the spectral measure now defined on the unit circle.When processing WSS random signals with linear, time-invariant (LTI) filters, it is helpful to think of the correlation function as a linear operator. Since it is a circulant operator (depends only on the difference between the two arguments), its eigenfunctions are the Fourier complex exponentials. Additionally, since the eigenfunctions of LTI operators are also complex exponentials, LTI processing of WSS random signals is highly tractable—all computations can be performed in the frequency domain. Thus, the WSS assumption is widely employed in signal processing algorithms.

Definition for complex stochastic process

In the case where left{X_tright} is a complex stochastic process the autocovariance function is defined as K_{XX}(t_1, t_2) = operatorname E[(X_{t_1}-m_X(t_1))overline{(X_{t_2}-m_X(t_2))}] and, in addition to the requirements in {{EquationNote|Eq.3}}, it is required that the pseudo-autocovariance function J_{XX}(t_1, t_2) = operatorname E[(X_{t_1}-m_X(t_1))(X_{t_2}-m_X(t_2))] depends only on the time lag. In formulas, left{X_tright} is WSS, if{{Equation box 1|indent =|title=
< infty & & text{for all } t in mathbb{R}end{align}
|begin{align}& m_X(t) = m_X(t + tau) & & text{for all } tau,t in mathbb{R} & K_{XX}(t_1, t_2) = K_{XX}(t_1 - t_2, 0) & & text{for all } t_1,t_2 in mathbb{R} & J_{XX}(t_1, t_2) = J_{XX}(t_1 - t_2, 0) & & text{for all } t_1,t_2 in mathbb{R} & operatorname E[|X(t)|^2] Eq.4}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}

Joint stationarity

The concept of stationarity may be extended to two stochastic processes.

Joint strict-sense stationarity

Two stochastic processes left{X_tright} and left{Y_tright} are called jointly strict-sense stationary if their joint cumulative distribution F_{XY}(x_{t_1} ,ldots, x_{t_m},y_{t_1^'} ,ldots, y_{t_n^'}) remains unchanged under time shifts, i.e. if{{Equation box 1|indent =|title=
|F_{XY}(x_{t_1} ,ldots, x_{t_m},y_{t_1^'} ,ldots, y_{t_n^'}) = F_{XY}(x_{t_1+tau} ,ldots, x_{t_m+tau},y_{t_1^'+tau} ,ldots, y_{t_n^'+tau})
quad text{for all } tau,t_1, ldots, t_m, t_1^', ldots, t_n^' in mathbb{R} text{ and for all } m,n in mathbb{N}|{{EquationRef|Eq.5}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}

Joint (M + N)th-order stationarity

Two random processes left{X_tright} and left{Y_tright} is said to be jointly (M + N)-th-order stationary if:{{rp|p. 159}}{{Equation box 1|indent =|title=
|F_{XY}(x_{t_1} ,ldots, x_{t_m},y_{t_1^'} ,ldots, y_{t_n^'}) = F_{XY}(x_{t_1+tau} ,ldots, x_{t_m+tau},y_{t_1^'+tau} ,ldots, y_{t_n^'+tau})
quad text{for all } tau,t_1, ldots, t_m, t_1^', ldots, t_n^' in mathbb{R} text{ and for all } m in {1,ldots,M}, n in {1,ldots,N}|{{EquationRef|Eq.6}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}

Joint weak or wide-sense stationarity

Two stochastic processes left{X_tright} and left{Y_tright} are called jointly wide-sense stationary if they are both wide-sense stationary and their cross-covariance function K_{XY}(t_1, t_2) = operatorname E[(X_{t_1}-m_X(t_1))(Y_{t_2}-m_Y(t_2))] depends only on the time difference tau = t_1 - t_2. This may be summarized as follows:{{Equation box 1|indent =|title=
|begin{align}& m_X(t) = m_X(t + tau) & & text{for all } tau,t in mathbb{R} & m_Y(t) = m_Y(t + tau) & & text{for all } tau,t in mathbb{R} & K_{XX}(t_1, t_2) = K_{XX}(t_1 - t_2, 0) & & text{for all } t_1,t_2 in mathbb{R} & K_{YY}(t_1, t_2) = K_{YY}(t_1 - t_2, 0) & & text{for all } t_1,t_2 in mathbb{R} & K_{XY}(t_1, t_2) = K_{XY}(t_1 - t_2, 0) & & text{for all } t_1,t_2 in mathbb{R}end{align}Eq.7}}}}|cellpadding= 6|border|border colour = #0073CF|background colour=#F5FFFA}}

Relation between types of stationarity

  • If a stochastic process is N-th-order stationary, then it is also M-th-order stationary for all {{tmath|M le N}}.
  • If a stochastic process is second order stationary (N=2) and has finite second moments, then it is also wide-sense stationary.{{rp|p. 159}}
  • If a stochastic process is wide-sense stationary, it is not necessarily second-order stationary.{{rp|p. 159}}
  • If a stochastic process is strict-sense stationary and has finite second moments, it is wide-sense stationary.{{rp|p. 299}}
  • If two stochastic processes are jointly (M + N)-th-order stationary, this does not guarantee that the individual processes are M-th- respectively N-th-order stationary.{{rp|p. 159}}

Other terminology

The terminology used for types of stationarity other than strict stationarity can be rather mixed. Some examples follow.
  • Priestley uses stationary up to order m if conditions similar to those given here for wide sense stationarity apply relating to moments up to order m.BOOK, Priestley, M. B., 1981, Spectral Analysis and Time Series, Academic Press, 0-12-564922-3, BOOK, Priestley, M. B., 1988, Non-linear and Non-stationary Time Series Analysis,weblink registration, Academic Press, 0-12-564911-8, Thus wide sense stationarity would be equivalent to "stationary to order 2", which is different from the definition of second-order stationarity given here.
  • Honarkhah and Caers also use the assumption of stationarity in the context of multiple-point geostatistics, where higher n-point statistics are assumed to be stationary in the spatial domain.JOURNAL, Honarkhah, M., Caers, J., 2010, 10.1007/s11004-010-9276-7, Stochastic Simulation of Patterns Using Distance-Based Pattern Modeling, Mathematical Geosciences, 42, 5, 487–517,
  • Tahmasebi and Sahimi have presented an adaptive Shannon-based methodology that can be used for modeling of any non-stationary systems.JOURNAL, Tahmasebi, P., Sahimi, M., 2015, 10.1103/PhysRevE.91.032401, Reconstruction of nonstationary disordered materials and media: Watershed transform and cross-correlation function,weblink PDF, Physical Review E, 91, 3, 25871117, 032401, free,

Differencing

One way to make some time series stationary is to compute the differences between consecutive observations. This is known as differencing. Differencing can help stabilize the mean of a time series by removing changes in the level of a time series, and so eliminating trends. This can also remove seasonality, if differences are taken appropriately (e.g. differencing observations 1 year apart to remove year-lo).Transformations such as logarithms can help to stabilize the variance of a time series.One of the ways for identifying non-stationary times series is the ACF plot. Sometimes, patterns will be more visible in the ACF plot than in the original time series; however, this is not always the case.WEB,weblink 8.1 Stationarity and differencing {{!, OTexts|website=www.otexts.org|access-date=2016-05-18}} Another approach to identifying non-stationarity is to look at the Laplace transform of a series, which will identify both exponential trends and sinusoidal seasonality (complex exponential trends). Related techniques from signal analysis such as the wavelet transform and Fourier transform may also be helpful.

See also

References

{{Reflist}}

Further reading

  • BOOK, Enders, Walter, Applied Econometric Time Series, New York, Wiley, 2010, Third, 978-0-470-50539-7, 53–57,
  • JOURNAL, Jestrovic, I., Coyle, J. L., Sejdic, E, 2015, 10.1016/j.brainres.2014.09.035, The effects of increased fluid viscosity on stationary characteristics of EEG signal in healthy adults, Brain Research, 1589, 45–53, 25245522, 4253861,
  • Hyndman, Athanasopoulos (2013). Forecasting: Principles and Practice. Otexts.weblink

External links

{{Stochastic processes}}{{Statistics|analysis}}


- content above as imported from Wikipedia
- "stationary process" does not exist on GetWiki (yet)
- time: 4:02am EDT - Sat, May 18 2024
[ this remote article is provided by Wikipedia ]
LATEST EDITS [ see all ]
GETWIKI 23 MAY 2022
GETWIKI 09 JUL 2019
Eastern Philosophy
History of Philosophy
GETWIKI 09 MAY 2016
GETWIKI 18 OCT 2015
M.R.M. Parrott
Biographies
GETWIKI 20 AUG 2014
CONNECT