World Library  
Flag as Inappropriate
Email this Article

Stationary process

Article Id: WHEBN0000329898
Reproduction Date:

Title: Stationary process  
Author: World Heritage Encyclopedia
Language: English
Subject: Autoregressive–moving-average model, Hilbert–Huang transform, Autocorrelation, Time series, Stationary distribution
Collection: Signal Processing, Stochastic Processes
Publisher: World Heritage Encyclopedia

Stationary process

In mathematics and statistics, a stationary process (or strict(ly) stationary process or strong(ly) stationary process) is a stochastic process whose joint probability distribution does not change when shifted in time. Consequently, parameters such as the mean and variance, if they are present, also do not change over time and do not follow any trends.

Stationarity is used as a tool in time series analysis, where the raw data is often transformed to become stationary; for example, economic data are often seasonal and/or dependent on a non-stationary price level. An important type of non-stationary process that does not include a trend-like behavior is the cyclostationary process.

Note that a "stationary process" is not the same thing as a "process with a stationary distribution". Indeed, there are further possibilities for confusion with the use of "stationary" in the context of stochastic processes; for example a "time-homogeneous" Markov chain is sometimes said to have "stationary transition probabilities". Besides, all stationary Markov random processes are time-homogeneous.


  • Definition 1
  • Examples 2
  • Weaker forms of stationarity 3
    • Weak or wide-sense stationarity 3.1
    • Other terminology 3.2
  • See also 4
  • References 5
  • Further reading 6
  • External links 7


Formally, let \left\{X_t\right\} be a stochastic process and let F_{X}(x_{t_1 + \tau}, \ldots, x_{t_k + \tau}) represent the cumulative distribution function of the joint distribution of \left\{X_t\right\} at times t_1 + \tau, \ldots, t_k + \tau. Then, \left\{X_t\right\} is said to be strictly(or strongly) stationary if, for all k, for all \tau, and for all t_1, \ldots, t_k,

F_{X}(x_{t_1+\tau} ,\ldots, x_{t_k+\tau}) = F_{X}(x_{t_1},\ldots, x_{t_k}).

Since \tau does not affect F_X(\cdot), F_{X} is not a function of time.


Two simulated time series processes, one stationary the other non-stationary. The Augmented Dickey–Fuller test is reported for each process and non-stationarity cannot be rejected for the second process.

As an example, white noise is stationary. The sound of a cymbal clashing, if hit only once, is not stationary because the acoustic power of the clash (and hence its variance) diminishes with time. However, it would be possible to invent a stochastic process describing when the cymbal is hit, such that the overall response would form a stationary process. For example, if the cymbal were hit at moments in time corresponding to a homogeneous Poisson Process, the overall response would be stationary.

An example of a discrete-time stationary process where the sample space is also discrete (so that the random variable may take one of N possible values) is a Bernoulli scheme. Other examples of a discrete-time stationary process with continuous sample space include some autoregressive and moving average processes which are both subsets of the autoregressive moving average model. Models with a non-trivial autoregressive component may be either stationary or non-stationary, depending on the parameter values, and important non-stationary special cases are where unit roots exist in the model.

Let Y be any scalar random variable, and define a time-series { Xt }, by

X_t=Y \qquad \text{ for all } t..

Then { Xt } is a stationary time series, for which realisations consist of a series of constant values, with a different constant value for each realisation. A law of large numbers does not apply on this case, as the limiting value of an average from a single realisation takes the random value determined by Y, rather than taking the expected value of Y.

As a further example of a stationary process for which any single realisation has an apparently noise-free structure, let Y have a uniform distribution on (0,2π] and define the time series { Xt } by

X_t=\cos (t+Y) \quad \text{ for } t \in \mathbb{R}.

Then { Xt } is strictly stationary.

Weaker forms of stationarity

Weak or wide-sense stationarity

A weaker form of stationarity commonly employed in signal processing is known as weak-sense stationarity, wide-sense stationarity (WSS), covariance stationarity, or second-order stationarity. WSS random processes only require that 1st moment and autocovariance do not vary with respect to time. Any strictly stationary process which has a mean and a covariance is also WSS.

So, a continuous-time random process x(t) which is WSS has the following restrictions on its mean function

\mathbb{E}[x(t)] = m_x(t) = m_x(t + \tau) \,\, \text{ for all } \, \tau \in \mathbb{R}

and autocovariance function

\mathbb{E}[(x(t_1)-m_x(t_1))(x(t_2)-m_x(t_2))] = C_x(t_1, t_2) = C_x(t_1 + (-t_2), t_2 + (-t_2)) = C_x(t_1 - t_2, 0).

The first property implies that the mean function mx(t) must be constant. The second property implies that the covariance function depends only on the difference between t_1 and t_2 and only needs to be indexed by one variable rather than two variables. Thus, instead of writing,

\,\!C_x(t_1 - t_2, 0)\,

the notation is often abbreviated and written as:

C_x(\tau) \,\! \mbox{ where } \tau = t_1 - t_2.

This also implies that the autocorrelation depends only on \tau = t_1 - t_2, that is

\,\! R_x(t_1,t_2) = R_x(t_1-t_2).

The main advantage of wide-sense stationarity is that it places the time-series in the context of Hilbert spaces. Let H be the Hilbert space generated by {x(t)} (that is, the closure of the set of all linear combinations of these random variables in the Hilbert space of all square-integrable random variables on the given probability space). By the positive definiteness of the autocovariance function, it follows from Bochner's theorem that there exists a positive measure μ on the real line such that H is isomorphic to the Hilbert subspace of L2(μ) generated by {e−2πiξ⋅t}. This then gives the following Fourier-type decomposition for continuous time stationary stochastic process: there exists a stochastic process ωξ with orthogonal increments such that, for all t

x(t) = \int e^{- 2 \pi i \lambda \cdot t} d \omega_{\lambda},

where the integral on the right hand side is interpreted in a suitable (Riemann) sense. Same result holds for a discrete-time stationary process, with the spectral measure now defined on the unit circle.

When processing WSS random signals with linear, time-invariant (LTI) filters, it is helpful to think of the correlation function as a linear operator. Since it is a circulant operator (depends only on the difference between the two arguments), its eigenfunctions are the Fourier complex exponentials. Additionally, since the eigenfunctions of LTI operators are also complex exponentials, LTI processing of WSS random signals is highly tractable—all computations can be performed in the frequency domain. Thus, the WSS assumption is widely employed in signal processing algorithms.

Other terminology

The terminology used for types of stationarity other than strict stationarity can be rather mixed. Some examples follow.

  • Priestley uses stationary up to order m if conditions similar to those given here for wide sense stationarity apply relating to moments up to order m.[1][2] Thus wide sense stationarity would be equivalent to "stationary to order 2", which is different from the definition of second-order stationarity given here.
  • Honarkhah and Caers also use the assumption of stationarity in the context of multiple-point geostatistics, where higher n-point statistics are assumed to be stationary in the spatial domain.[3]

See also


  1. ^ Priestley, M. B. (1981). Spectral Analysis and Time Series. Academic Press.  
  2. ^ Priestley, M. B. (1988). Non-linear and Non-stationary Time Series Analysis. Academic Press.  
  3. ^ Honarkhah, M.; Caers, J. (2010). "Stochastic Simulation of Patterns Using Distance-Based Pattern Modeling". Mathematical Geosciences 42 (5): 487–517.  

Further reading

  • Enders, Walter (2010). Applied Econometric Time Series (Third ed.). New York: Wiley. pp. 53–57.  

External links

  • Spectral decomposition of a random function (Springer)
This article was sourced from Creative Commons Attribution-ShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for and content contributors is made possible from the U.S. Congress, E-Government Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a non-profit organization.

Copyright © World Library Foundation. All rights reserved. eBooks from World Library are sponsored by the World Library Foundation,
a 501c(4) Member's Support Non-Profit Organization, and is NOT affiliated with any governmental agency or department.