fbpx
Wikipedia

Stochastic process

In probability theory and related fields, a stochastic (/stəˈkæstɪk/) or random process is a mathematical object usually defined as a sequence of random variables in a probability space, where the index of the sequence often has the interpretation of time. Stochastic processes are widely used as mathematical models of systems and phenomena that appear to vary in a random manner. Examples include the growth of a bacterial population, an electrical current fluctuating due to thermal noise, or the movement of a gas molecule.[1][4][5] Stochastic processes have applications in many disciplines such as biology,[6] chemistry,[7] ecology,[8] neuroscience,[9] physics,[10] image processing, signal processing,[11] control theory,[12] information theory,[13] computer science,[14] and telecommunications.[15] Furthermore, seemingly random changes in financial markets have motivated the extensive use of stochastic processes in finance.[16][17][18]

A computer-simulated realization of a Wiener or Brownian motion process on the surface of a sphere. The Wiener process is widely considered the most studied and central stochastic process in probability theory.[1][2][3]

Applications and the study of phenomena have in turn inspired the proposal of new stochastic processes. Examples of such stochastic processes include the Wiener process or Brownian motion process,[a] used by Louis Bachelier to study price changes on the Paris Bourse,[21] and the Poisson process, used by A. K. Erlang to study the number of phone calls occurring in a certain period of time.[22] These two stochastic processes are considered the most important and central in the theory of stochastic processes,[1][4][23] and were invented repeatedly and independently, both before and after Bachelier and Erlang, in different settings and countries.[21][24]

The term random function is also used to refer to a stochastic or random process,[25][26] because a stochastic process can also be interpreted as a random element in a function space.[27][28] The terms stochastic process and random process are used interchangeably, often with no specific mathematical space for the set that indexes the random variables.[27][29] But often these two terms are used when the random variables are indexed by the integers or an interval of the real line.[5][29] If the random variables are indexed by the Cartesian plane or some higher-dimensional Euclidean space, then the collection of random variables is usually called a random field instead.[5][30] The values of a stochastic process are not always numbers and can be vectors or other mathematical objects.[5][28]

Based on their mathematical properties, stochastic processes can be grouped into various categories, which include random walks,[31] martingales,[32] Markov processes,[33] Lévy processes,[34] Gaussian processes,[35] random fields,[36] renewal processes, and branching processes.[37] The study of stochastic processes uses mathematical knowledge and techniques from probability, calculus, linear algebra, set theory, and topology[38][39][40] as well as branches of mathematical analysis such as real analysis, measure theory, Fourier analysis, and functional analysis.[41][42][43] The theory of stochastic processes is considered to be an important contribution to mathematics[44] and it continues to be an active topic of research for both theoretical reasons and applications.[45][46][47]

Introduction edit

A stochastic or random process can be defined as a collection of random variables that is indexed by some mathematical set, meaning that each random variable of the stochastic process is uniquely associated with an element in the set.[4][5] The set used to index the random variables is called the index set. Historically, the index set was some subset of the real line, such as the natural numbers, giving the index set the interpretation of time.[1] Each random variable in the collection takes values from the same mathematical space known as the state space. This state space can be, for example, the integers, the real line or  -dimensional Euclidean space.[1][5] An increment is the amount that a stochastic process changes between two index values, often interpreted as two points in time.[48][49] A stochastic process can have many outcomes, due to its randomness, and a single outcome of a stochastic process is called, among other names, a sample function or realization.[28][50]

 
A single computer-simulated sample function or realization, among other terms, of a three-dimensional Wiener or Brownian motion process for time 0 ≤ t ≤ 2. The index set of this stochastic process is the non-negative numbers, while its state space is three-dimensional Euclidean space.

Classifications edit

A stochastic process can be classified in different ways, for example, by its state space, its index set, or the dependence among the random variables. One common way of classification is by the cardinality of the index set and the state space.[51][52][53]

When interpreted as time, if the index set of a stochastic process has a finite or countable number of elements, such as a finite set of numbers, the set of integers, or the natural numbers, then the stochastic process is said to be in discrete time.[54][55] If the index set is some interval of the real line, then time is said to be continuous. The two types of stochastic processes are respectively referred to as discrete-time and continuous-time stochastic processes.[48][56][57] Discrete-time stochastic processes are considered easier to study because continuous-time processes require more advanced mathematical techniques and knowledge, particularly due to the index set being uncountable.[58][59] If the index set is the integers, or some subset of them, then the stochastic process can also be called a random sequence.[55]

If the state space is the integers or natural numbers, then the stochastic process is called a discrete or integer-valued stochastic process. If the state space is the real line, then the stochastic process is referred to as a real-valued stochastic process or a process with continuous state space. If the state space is  -dimensional Euclidean space, then the stochastic process is called a  -dimensional vector process or  -vector process.[51][52]

Etymology edit

The word stochastic in English was originally used as an adjective with the definition "pertaining to conjecturing", and stemming from a Greek word meaning "to aim at a mark, guess", and the Oxford English Dictionary gives the year 1662 as its earliest occurrence.[60] In his work on probability Ars Conjectandi, originally published in Latin in 1713, Jakob Bernoulli used the phrase "Ars Conjectandi sive Stochastice", which has been translated to "the art of conjecturing or stochastics".[61] This phrase was used, with reference to Bernoulli, by Ladislaus Bortkiewicz[62] who in 1917 wrote in German the word stochastik with a sense meaning random. The term stochastic process first appeared in English in a 1934 paper by Joseph Doob.[60] For the term and a specific mathematical definition, Doob cited another 1934 paper, where the term stochastischer Prozeß was used in German by Aleksandr Khinchin,[63][64] though the German term had been used earlier, for example, by Andrei Kolmogorov in 1931.[65]

According to the Oxford English Dictionary, early occurrences of the word random in English with its current meaning, which relates to chance or luck, date back to the 16th century, while earlier recorded usages started in the 14th century as a noun meaning "impetuosity, great speed, force, or violence (in riding, running, striking, etc.)". The word itself comes from a Middle French word meaning "speed, haste", and it is probably derived from a French verb meaning "to run" or "to gallop". The first written appearance of the term random process pre-dates stochastic process, which the Oxford English Dictionary also gives as a synonym, and was used in an article by Francis Edgeworth published in 1888.[66]

Terminology edit

The definition of a stochastic process varies,[67] but a stochastic process is traditionally defined as a collection of random variables indexed by some set.[68][69] The terms random process and stochastic process are considered synonyms and are used interchangeably, without the index set being precisely specified.[27][29][30][70][71][72] Both "collection",[28][70] or "family" are used[4][73] while instead of "index set", sometimes the terms "parameter set"[28] or "parameter space"[30] are used.

The term random function is also used to refer to a stochastic or random process,[5][74][75] though sometimes it is only used when the stochastic process takes real values.[28][73] This term is also used when the index sets are mathematical spaces other than the real line,[5][76] while the terms stochastic process and random process are usually used when the index set is interpreted as time,[5][76][77] and other terms are used such as random field when the index set is  -dimensional Euclidean space   or a manifold.[5][28][30]

Notation edit

A stochastic process can be denoted, among other ways, by  ,[56]  ,[69]  [78]   or simply as  . Some authors mistakenly write   even though it is an abuse of function notation.[79] For example,   or   are used to refer to the random variable with the index  , and not the entire stochastic process.[78] If the index set is  , then one can write, for example,   to denote the stochastic process.[29]

Examples edit

Bernoulli process edit

One of the simplest stochastic processes is the Bernoulli process,[80] which is a sequence of independent and identically distributed (iid) random variables, where each random variable takes either the value one or zero, say one with probability   and zero with probability  . This process can be linked to an idealisation of repeatedly flipping a coin, where the probability of obtaining a head is taken to be   and its value is one, while the value of a tail is zero.[81] In other words, a Bernoulli process is a sequence of iid Bernoulli random variables,[82] where each idealised coin flip is an example of a Bernoulli trial.[83]

Random walk edit

Random walks are stochastic processes that are usually defined as sums of iid random variables or random vectors in Euclidean space, so they are processes that change in discrete time.[84][85][86][87][88] But some also use the term to refer to processes that change in continuous time,[89] particularly the Wiener process used in financial models, which has led to some confusion, resulting in its criticism.[90] There are other various types of random walks, defined so their state spaces can be other mathematical objects, such as lattices and groups, and in general they are highly studied and have many applications in different disciplines.[89][91]

A classic example of a random walk is known as the simple random walk, which is a stochastic process in discrete time with the integers as the state space, and is based on a Bernoulli process, where each Bernoulli variable takes either the value positive one or negative one. In other words, the simple random walk takes place on the integers, and its value increases by one with probability, say,  , or decreases by one with probability  , so the index set of this random walk is the natural numbers, while its state space is the integers. If  , this random walk is called a symmetric random walk.[92][93]

Wiener process edit

The Wiener process is a stochastic process with stationary and independent increments that are normally distributed based on the size of the increments.[2][94] The Wiener process is named after Norbert Wiener, who proved its mathematical existence, but the process is also called the Brownian motion process or just Brownian motion due to its historical connection as a model for Brownian movement in liquids.[95][96][97]

 
Realizations of Wiener processes (or Brownian motion processes) with drift (blue) and without drift (red).

Playing a central role in the theory of probability, the Wiener process is often considered the most important and studied stochastic process, with connections to other stochastic processes.[1][2][3][98][99][100][101] Its index set and state space are the non-negative numbers and real numbers, respectively, so it has both continuous index set and states space.[102] But the process can be defined more generally so its state space can be  -dimensional Euclidean space.[91][99][103] If the mean of any increment is zero, then the resulting Wiener or Brownian motion process is said to have zero drift. If the mean of the increment for any two points in time is equal to the time difference multiplied by some constant  , which is a real number, then the resulting stochastic process is said to have drift  .[104][105][106]

Almost surely, a sample path of a Wiener process is continuous everywhere but nowhere differentiable. It can be considered as a continuous version of the simple random walk.[49][105] The process arises as the mathematical limit of other stochastic processes such as certain random walks rescaled,[107][108] which is the subject of Donsker's theorem or invariance principle, also known as the functional central limit theorem.[109][110][111]

The Wiener process is a member of some important families of stochastic processes, including Markov processes, Lévy processes and Gaussian processes.[2][49] The process also has many applications and is the main stochastic process used in stochastic calculus.[112][113] It plays a central role in quantitative finance,[114][115] where it is used, for example, in the Black–Scholes–Merton model.[116] The process is also used in different fields, including the majority of natural sciences as well as some branches of social sciences, as a mathematical model for various random phenomena.[3][117][118]

Poisson process edit

The Poisson process is a stochastic process that has different forms and definitions.[119][120] It can be defined as a counting process, which is a stochastic process that represents the random number of points or events up to some time. The number of points of the process that are located in the interval from zero to some given time is a Poisson random variable that depends on that time and some parameter. This process has the natural numbers as its state space and the non-negative numbers as its index set. This process is also called the Poisson counting process, since it can be interpreted as an example of a counting process.[119]

If a Poisson process is defined with a single positive constant, then the process is called a homogeneous Poisson process.[119][121] The homogeneous Poisson process is a member of important classes of stochastic processes such as Markov processes and Lévy processes.[49]

The homogeneous Poisson process can be defined and generalized in different ways. It can be defined such that its index set is the real line, and this stochastic process is also called the stationary Poisson process.[122][123] If the parameter constant of the Poisson process is replaced with some non-negative integrable function of  , the resulting process is called an inhomogeneous or nonhomogeneous Poisson process, where the average density of points of the process is no longer constant.[124] Serving as a fundamental process in queueing theory, the Poisson process is an important process for mathematical models, where it finds applications for models of events randomly occurring in certain time windows.[125][126]

Defined on the real line, the Poisson process can be interpreted as a stochastic process,[49][127] among other random objects.[128][129] But then it can be defined on the  -dimensional Euclidean space or other mathematical spaces,[130] where it is often interpreted as a random set or a random counting measure, instead of a stochastic process.[128][129] In this setting, the Poisson process, also called the Poisson point process, is one of the most important objects in probability theory, both for applications and theoretical reasons.[22][131] But it has been remarked that the Poisson process does not receive as much attention as it should, partly due to it often being considered just on the real line, and not on other mathematical spaces.[131][132]

Definitions edit

Stochastic process edit

A stochastic process is defined as a collection of random variables defined on a common probability space  , where   is a sample space,   is a  -algebra, and   is a probability measure; and the random variables, indexed by some set  , all take values in the same mathematical space  , which must be measurable with respect to some  -algebra  .[28]

In other words, for a given probability space   and a measurable space  , a stochastic process is a collection of  -valued random variables, which can be written as:[80]

 

Historically, in many problems from the natural sciences a point   had the meaning of time, so   is a random variable representing a value observed at time  .[133] A stochastic process can also be written as   to reflect that it is actually a function of two variables,   and  .[28][134]

There are other ways to consider a stochastic process, with the above definition being considered the traditional one.[68][69] For example, a stochastic process can be interpreted or defined as a  -valued random variable, where   is the space of all the possible functions from the set   into the space  .[27][68] However this alternative definition as a "function-valued random variable" in general requires additional regularity assumptions to be well-defined.[135]

Index set edit

The set   is called the index set[4][51] or parameter set[28][136] of the stochastic process. Often this set is some subset of the real line, such as the natural numbers or an interval, giving the set   the interpretation of time.[1] In addition to these sets, the index set   can be another set with a total order or a more general set,[1][54] such as the Cartesian plane   or  -dimensional Euclidean space, where an element   can represent a point in space.[48][137] That said, many results and theorems are only possible for stochastic processes with a totally ordered index set.[138]

State space edit

The mathematical space   of a stochastic process is called its state space. This mathematical space can be defined using integers, real lines,  -dimensional Euclidean spaces, complex planes, or more abstract mathematical spaces. The state space is defined using elements that reflect the different values that the stochastic process can take.[1][5][28][51][56]

Sample function edit

A sample function is a single outcome of a stochastic process, so it is formed by taking a single possible value of each random variable of the stochastic process.[28][139] More precisely, if   is a stochastic process, then for any point  , the mapping

 

is called a sample function, a realization, or, particularly when   is interpreted as time, a sample path of the stochastic process  .[50] This means that for a fixed  , there exists a sample function that maps the index set   to the state space  .[28] Other names for a sample function of a stochastic process include trajectory, path function[140] or path.[141]

Increment edit

An increment of a stochastic process is the difference between two random variables of the same stochastic process. For a stochastic process with an index set that can be interpreted as time, an increment is how much the stochastic process changes over a certain time period. For example, if   is a stochastic process with state space   and index set  , then for any two non-negative numbers   and   such that  , the difference   is a  -valued random variable known as an increment.[48][49] When interested in the increments, often the state space   is the real line or the natural numbers, but it can be  -dimensional Euclidean space or more abstract spaces such as Banach spaces.[49]

Further definitions edit

Law edit

For a stochastic process   defined on the probability space  , the law of stochastic process   is defined as the image measure:

 

where   is a probability measure, the symbol   denotes function composition and   is the pre-image of the measurable function or, equivalently, the  -valued random variable  , where   is the space of all the possible  -valued functions of  , so the law of a stochastic process is a probability measure.[27][68][142][143]

For a measurable subset   of  , the pre-image of   gives

 

so the law of a   can be written as:[28]

 

The law of a stochastic process or a random variable is also called the probability law, probability distribution, or the distribution.[133][142][144][145][146]

Finite-dimensional probability distributions edit

For a stochastic process   with law  , its finite-dimensional distribution for   is defined as:

 

This measure  is the joint distribution of the random vector  ; it can be viewed as a "projection" of the law   onto a finite subset of  .[27][147]

For any measurable subset   of the  -fold Cartesian power  , the finite-dimensional distributions of a stochastic process   can be written as:[28]

 

The finite-dimensional distributions of a stochastic process satisfy two mathematical conditions known as consistency conditions.[57]

Stationarity edit

Stationarity is a mathematical property that a stochastic process has when all the random variables of that stochastic process are identically distributed. In other words, if   is a stationary stochastic process, then for any   the random variable   has the same distribution, which means that for any set of   index set values  , the corresponding   random variables

 

all have the same probability distribution. The index set of a stationary stochastic process is usually interpreted as time, so it can be the integers or the real line.[148][149] But the concept of stationarity also exists for point processes and random fields, where the index set is not interpreted as time.[148][150][151]

When the index set   can be interpreted as time, a stochastic process is said to be stationary if its finite-dimensional distributions are invariant under translations of time. This type of stochastic process can be used to describe a physical system that is in steady state, but still experiences random fluctuations.[148] The intuition behind stationarity is that as time passes the distribution of the stationary stochastic process remains the same.[152] A sequence of random variables forms a stationary stochastic process only if the random variables are identically distributed.[148]

A stochastic process with the above definition of stationarity is sometimes said to be strictly stationary, but there are other forms of stationarity. One example is when a discrete-time or continuous-time stochastic process   is said to be stationary in the wide sense, then the process   has a finite second moment for all   and the covariance of the two random variables   and   depends only on the number   for all  .[152][153] Khinchin introduced the related concept of stationarity in the wide sense, which has other names including covariance stationarity or stationarity in the broad sense.[153][154]

Filtration edit

A filtration is an increasing sequence of sigma-algebras defined in relation to some probability space and an index set that has some total order relation, such as in the case of the index set being some subset of the real numbers. More formally, if a stochastic process has an index set with a total order, then a filtration  , on a probability space   is a family of sigma-algebras such that   for all  , where   and   denotes the total order of the index set  .[51] With the concept of a filtration, it is possible to study the amount of information contained in a stochastic process   at  , which can be interpreted as time  .[51][155] The intuition behind a filtration   is that as time   passes, more and more information on   is known or available, which is captured in  , resulting in finer and finer partitions of  .[156][157]

Modification edit

A modification of a stochastic process is another stochastic process, which is closely related to the original stochastic process. More precisely, a stochastic process   that has the same index set  , state space  , and probability space   as another stochastic process   is said to be a modification of   if for all   the following

 

holds. Two stochastic processes that are modifications of each other have the same finite-dimensional law[158] and they are said to be stochastically equivalent or equivalent.[159]

Instead of modification, the term version is also used,[150][160][161][162] however some authors use the term version when two stochastic processes have the same finite-dimensional distributions, but they may be defined on different probability spaces, so two processes that are modifications of each other, are also versions of each other, in the latter sense, but not the converse.[163][142]

If a continuous-time real-valued stochastic process meets certain moment conditions on its increments, then the Kolmogorov continuity theorem says that there exists a modification of this process that has continuous sample paths with probability one, so the stochastic process has a continuous modification or version.[161][162][164] The theorem can also be generalized to random fields so the index set is  -dimensional Euclidean space[165] as well as to stochastic processes with metric spaces as their state spaces.[166]

Indistinguishable edit

Two stochastic processes   and   defined on the same probability space   with the same index set   and set space   are said be indistinguishable if the following

 

holds.[142][158] If two   and   are modifications of each other and are almost surely continuous, then   and   are indistinguishable.[167]

Separability edit

Separability is a property of a stochastic process based on its index set in relation to the probability measure. The property is assumed so that functionals of stochastic processes or random fields with uncountable index sets can form random variables. For a stochastic process to be separable, in addition to other conditions, its index set must be a separable space,[b] which means that the index set has a dense countable subset.[150][168]

More precisely, a real-valued continuous-time stochastic process   with a probability space   is separable if its index set   has a dense countable subset   and there is a set   of probability zero, so  , such that for every open set   and every closed set  , the two events   and   differ from each other at most on a subset of  .[169][170][171] The definition of separability[c] can also be stated for other index sets and state spaces,[174] such as in the case of random fields, where the index set as well as the state space can be  -dimensional Euclidean space.[30][150]

The concept of separability of a stochastic process was introduced by Joseph Doob,.[168] The underlying idea of separability is to make a countable set of points of the index set determine the properties of the stochastic process.[172] Any stochastic process with a countable index set already meets the separability conditions, so discrete-time stochastic processes are always separable.[175] A theorem by Doob, sometimes known as Doob's separability theorem, says that any real-valued continuous-time stochastic process has a separable modification.[168][170][176] Versions of this theorem also exist for more general stochastic processes with index sets and state spaces other than the real line.[136]

Independence edit

Two stochastic processes   and   defined on the same probability space   with the same index set   are said be independent if for all   and for every choice of epochs  , the random vectors   and   are independent.[177]: p. 515 

Uncorrelatedness edit

Two stochastic processes   and   are called uncorrelated if their cross-covariance   is zero for all times.[178]: p. 142  Formally:

 .

Independence implies uncorrelatedness edit

If two stochastic processes   and   are independent, then they are also uncorrelated.[178]: p. 151 

Orthogonality edit

Two stochastic processes   and   are called orthogonal if their cross-correlation   is zero for all times.[178]: p. 142  Formally:

 .

Skorokhod space edit

A Skorokhod space, also written as Skorohod space, is a mathematical space of all the functions that are right-continuous with left limits, defined on some interval of the real line such as   or  , and take values on the real line or on some metric space.[179][180][181] Such functions are known as càdlàg or cadlag functions, based on the acronym of the French phrase continue à droite, limite à gauche.[179][182] A Skorokhod function space, introduced by Anatoliy Skorokhod,[181] is often denoted with the letter  ,[179][180][181][182] so the function space is also referred to as space  .[179][183][184] The notation of this function space can also include the interval on which all the càdlàg functions are defined, so, for example,   denotes the space of càdlàg functions defined on the unit interval  .[182][184][185]

Skorokhod function spaces are frequently used in the theory of stochastic processes because it often assumed that the sample functions of continuous-time stochastic processes belong to a Skorokhod space.[181][183] Such spaces contain continuous functions, which correspond to sample functions of the Wiener process. But the space also has functions with discontinuities, which means that the sample functions of stochastic processes with jumps, such as the Poisson process (on the real line), are also members of this space.[184][186]

Regularity edit

In the context of mathematical construction of stochastic processes, the term regularity is used when discussing and assuming certain conditions for a stochastic process to resolve possible construction issues.[187][188] For example, to study stochastic processes with uncountable index sets, it is assumed that the stochastic process adheres to some type of regularity condition such as the sample functions being continuous.[189][190]

Further examples edit

Markov processes and chains edit

Markov processes are stochastic processes, traditionally in discrete or continuous time, that have the Markov property, which means the next value of the Markov process depends on the current value, but it is conditionally independent of the previous values of the stochastic process. In other words, the behavior of the process in the future is stochastically independent of its behavior in the past, given the current state of the process.[191][192]

The Brownian motion process and the Poisson process (in one dimension) are both examples of Markov processes[193] in continuous time, while random walks on the integers and the gambler's ruin problem are examples of Markov processes in discrete time.[194][195]

A Markov chain is a type of Markov process that has either discrete state space or discrete index set (often representing time), but the precise definition of a Markov chain varies.[196] For example, it is common to define a Markov chain as a Markov process in either discrete or continuous time with a countable state space (thus regardless of the nature of time),[197][198][199][200] but it has been also common to define a Markov chain as having discrete time in either countable or continuous state space (thus regardless of the state space).[196] It has been argued that the first definition of a Markov chain, where it has discrete time, now tends to be used, despite the second definition having been used by researchers like Joseph Doob and Kai Lai Chung.[201]

Markov processes form an important class of stochastic processes and have applications in many areas.[39][202] For example, they are the basis for a general stochastic simulation method known as Markov chain Monte Carlo, which is used for simulating random objects with specific probability distributions, and has found application in Bayesian statistics.[203][204]

The concept of the Markov property was originally for stochastic processes in continuous and discrete time, but the property has been adapted for other index sets such as  -dimensional Euclidean space, which results in collections of random variables known as Markov random fields.[205][206][207]

Martingale edit

A martingale is a discrete-time or continuous-time stochastic process with the property that, at every instant, given the current value and all the past values of the process, the conditional expectation of every future value is equal to the current value. In discrete time, if this property holds for the next value, then it holds for all future values. The exact mathematical definition of a martingale requires two other conditions coupled with the mathematical concept of a filtration, which is related to the intuition of increasing available information as time passes. Martingales are usually defined to be real-valued,[208][209][155] but they can also be complex-valued[210] or even more general.[211]

A symmetric random walk and a Wiener process (with zero drift) are both examples of martingales, respectively, in discrete and continuous time.[208][209] For a sequence of independent and identically distributed random variables   with zero mean, the stochastic process formed from the successive partial sums   is a discrete-time martingale.[212] In this aspect, discrete-time martingales generalize the idea of partial sums of independent random variables.[213]

Martingales can also be created from stochastic processes by applying some suitable transformations, which is the case for the homogeneous Poisson process (on the real line) resulting in a martingale called the compensated Poisson process.[209] Martingales can also be built from other martingales.[212] For example, there are martingales based on the martingale the Wiener process, forming continuous-time martingales.[208][214]

Martingales mathematically formalize the idea of a 'fair game' where it is possible form reasonable expectations for payoffs,[215] and they were originally developed to show that it is not possible to gain an 'unfair' advantage in such a game.[216] But now they are used in many areas of probability, which is one of the main reasons for studying them.[155][216][217] Many problems in probability have been solved by finding a martingale in the problem and studying it.[218] Martingales will converge, given some conditions on their moments, so they are often used to derive convergence results, due largely to martingale convergence theorems.[213][219][220]

Martingales have many applications in statistics, but it has been remarked that its use and application are not as widespread as it could be in the field of statistics, particularly statistical inference.[221] They have found applications in areas in probability theory such as queueing theory and Palm calculus[222] and other fields such as economics[223] and finance.[17]

Lévy process edit

Lévy processes are types of stochastic processes that can be considered as generalizations of random walks in continuous time.[49][224] These processes have many applications in fields such as finance, fluid mechanics, physics and biology.[225][226] The main defining characteristics of these processes are their stationarity and independence properties, so they were known as processes with stationary and independent increments. In other words, a stochastic process   is a Lévy process if for   non-negatives numbers,  , the corresponding   increments

 

are all independent of each other, and the distribution of each increment only depends on the difference in time.[49]

A Lévy process can be defined such that its state space is some abstract mathematical space, such as a Banach space, but the processes are often defined so that they take values in Euclidean space. The index set is the non-negative numbers, so  , which gives the interpretation of time. Important stochastic processes such as the Wiener process, the homogeneous Poisson process (in one dimension), and subordinators are all Lévy processes.[49][224]

Random field edit

A random field is a collection of random variables indexed by a  -dimensional Euclidean space or some manifold. In general, a random field can be considered an example of a stochastic or random process, where the index set is not necessarily a subset of the real line.[30] But there is a convention that an indexed collection of random variables is called a random field when the index has two or more dimensions.[5][28][227] If the specific definition of a stochastic process requires the index set to be a subset of the real line, then the random field can be considered as a generalization of stochastic process.[228]

Point process edit

A point process is a collection of points randomly located on some mathematical space such as the real line,  -dimensional Euclidean space, or more abstract spaces. Sometimes the term point process is not preferred, as historically the word process denoted an evolution of some system in time, so a point process is also called a random point field.[229] There are different interpretations of a point process, such a random counting measure or a random set.[230][231] Some authors regard a point process and stochastic process as two different objects such that a point process is a random object that arises from or is associated with a stochastic process,[232][233] though it has been remarked that the difference between point processes and stochastic processes is not clear.[233]

Other authors consider a point process as a stochastic process, where the process is indexed by sets of the underlying space[d] on which it is defined, such as the real line or  -dimensional Euclidean space.[236][237] Other stochastic processes such as renewal and counting processes are studied in the theory of point processes.[238][233]

History edit

Early probability theory edit

Probability theory has its origins in games of chance, which have a long history, with some games being played thousands of years ago,[239][240] but very little analysis on them was done in terms of probability.[239][241] The year 1654 is often considered the birth of probability theory when French mathematicians Pierre Fermat and Blaise Pascal had a written correspondence on probability, motivated by a gambling problem.[239][242][243] But there was earlier mathematical work done on the probability of gambling games such as Liber de Ludo Aleae by Gerolamo Cardano, written in the 16th century but posthumously published later in 1663.[239][244]

After Cardano, Jakob Bernoulli[e] wrote Ars Conjectandi, which is considered a significant event in the history of probability theory.[239] Bernoulli's book was published, also posthumously, in 1713 and inspired many mathematicians to study probability.[239][246][247] But despite some renowned mathematicians contributing to probability theory, such as Pierre-Simon Laplace, Abraham de Moivre, Carl Gauss, Siméon Poisson and Pafnuty Chebyshev,[248][249] most of the mathematical community[f] did not consider probability theory to be part of mathematics until the 20th century.[248][250][251][252]

Statistical mechanics edit

In the physical sciences, scientists developed in the 19th century the discipline of statistical mechanics, where physical systems, such as containers filled with gases, are regarded or treated mathematically as collections of many moving particles. Although there were attempts to incorporate randomness into statistical physics by some scientists, such as Rudolf Clausius, most of the work had little or no randomness.[253][254] This changed in 1859 when James Clerk Maxwell contributed significantly to the field, more specifically, to the kinetic theory of gases, by presenting work where he modelled the gas particles as moving in random directions at random velocities.[255][256] The kinetic theory of gases and statistical physics continued to be developed in the second half of the 19th century, with work done chiefly by Clausius, Ludwig Boltzmann and Josiah Gibbs, which would later have an influence on Albert Einstein's mathematical model for Brownian movement.[257]

Measure theory and probability theory edit

At the International Congress of Mathematicians in Paris in 1900, David Hilbert presented a list of mathematical problems, where his sixth problem asked for a mathematical treatment of physics and probability involving axioms.[249] Around the start of the 20th century, mathematicians developed measure theory, a branch of mathematics for studying integrals of mathematical functions, where two of the founders were French mathematicians, Henri Lebesgue and Émile Borel. In 1925, another French mathematician Paul Lévy published the first probability book that used ideas from measure theory.[249]

In the 1920s, fundamental contributions to probability theory were made in the Soviet Union by mathematicians such as Sergei Bernstein, Aleksandr Khinchin,[g] and Andrei Kolmogorov.[252] Kolmogorov published in 1929 his first attempt at presenting a mathematical foundation, based on measure theory, for probability theory.[258] In the early 1930s, Khinchin and Kolmogorov set up probability seminars, which were attended by researchers such as Eugene Slutsky and Nikolai Smirnov,[259] and Khinchin gave the first mathematical definition of a stochastic process as a set of random variables indexed by the real line.[63][260][h]

Birth of modern probability theory edit

In 1933, Andrei Kolmogorov published in German, his book on the foundations of probability theory titled Grundbegriffe der Wahrscheinlichkeitsrechnung,[i] where Kolmogorov used measure theory to develop an axiomatic framework for probability theory. The publication of this book is now widely considered to be the birth of modern probability theory, when the theories of probability and stochastic processes became parts of mathematics.[249][252]

After the publication of Kolmogorov's book, further fundamental work on probability theory and stochastic processes was done by Khinchin and Kolmogorov as well as other mathematicians such as Joseph Doob, William Feller, Maurice Fréchet, Paul Lévy, Wolfgang Doeblin, and Harald Cramér.[249][252] Decades later, Cramér referred to the 1930s as the "heroic period of mathematical probability theory".[252] World War II greatly interrupted the development of probability theory, causing, for example, the migration of Feller from Sweden to the United States of America[252] and the death of Doeblin, considered now a pioneer in stochastic processes.[262]

 
Mathematician Joseph Doob did early work on the theory of stochastic processes, making fundamental contributions, particularly in the theory of martingales.[263][261] His book Stochastic Processes is considered highly influential in the field of probability theory.[264]

Stochastic processes after World War II edit

After World War II, the study of probability theory and stochastic processes gained more attention from mathematicians, with significant contributions made in many areas of probability and mathematics as well as the creation of new areas.[252][265] Starting in the 1940s, Kiyosi Itô published papers developing the field of stochastic calculus, which involves stochastic integrals and stochastic differential equations based on the Wiener or Brownian motion process.[266]

Also starting in the 1940s, connections were made between stochastic processes, particularly martingales, and the mathematical field of potential theory, with early ideas by Shizuo Kakutani and then later work by Joseph Doob.[265] Further work, considered pioneering, was done by Gilbert Hunt in the 1950s, connecting Markov processes and potential theory, which had a significant effect on the theory of Lévy processes and led to more interest in studying Markov processes with methods developed by Itô.[21][267][268]

In 1953, Doob published his book Stochastic processes, which had a strong influence on the theory of stochastic processes and stressed the importance of measure theory in probability.[265] [264] Doob also chiefly developed the theory of martingales, with later substantial contributions by Paul-André Meyer. Earlier work had been carried out by Sergei Bernstein, Paul Lévy and Jean Ville, the latter adopting the term martingale for the stochastic process.[269][270] Methods from the theory of martingales became popular for solving various probability problems. Techniques and theory were developed to study Markov processes and then applied to martingales. Conversely, methods from the theory of martingales were established to treat Markov processes.[265]

Other fields of probability were developed and used to study stochastic processes, with one main approach being the theory of large deviations.[265] The theory has many applications in statistical physics, among other fields, and has core ideas going back to at least the 1930s. Later in the 1960s and 1970s, fundamental work was done by Alexander Wentzell in the Soviet Union and Monroe D. Donsker and Srinivasa Varadhan in the United States of America,[271] which would later result in Varadhan winning the 2007 Abel Prize.[272] In the 1990s and 2000s the theories of Schramm–Loewner evolution[273] and rough paths[142] were introduced and developed to study stochastic processes and other mathematical objects in probability theory, which respectively resulted in Fields Medals being awarded to Wendelin Werner[274] in 2008 and to Martin Hairer in 2014.[275]

The theory of stochastic processes still continues to be a focus of research, with yearly international conferences on the topic of stochastic processes.[45][225]

Discoveries of specific stochastic processes edit

Although Khinchin gave mathematical definitions of stochastic processes in the 1930s,[63][260] specific stochastic processes had already been discovered in different settings, such as the Brownian motion process and the Poisson process.[21][24] Some families of stochastic processes such as point processes or renewal processes have long and complex histories, stretching back centuries.[276]

Bernoulli process edit

The Bernoulli process, which can serve as a mathematical model for flipping a biased coin, is possibly the first stochastic process to have been studied.[81] The process is a sequence of independent Bernoulli trials,[82] which are named after Jackob Bernoulli who used them to study games of chance, including probability problems proposed and studied earlier by Christiaan Huygens.[277] Bernoulli's work, including the Bernoulli process, were published in his book Ars Conjectandi in 1713.[278]

Random walks edit

In 1905, Karl Pearson coined the term random walk while posing a problem describing a random walk on the plane, which was motivated by an application in biology, but such problems involving random walks had already been studied in other fields. Certain gambling problems that were studied centuries earlier can be considered as problems involving random walks.[89][278] For example, the problem known as the Gambler's ruin is based on a simple random walk,[195][279] and is an example of a random walk with absorbing barriers.[242][280] Pascal, Fermat and Huyens all gave numerical solutions to this problem without detailing their methods,[281] and then more detailed solutions were presented by Jakob Bernoulli and Abraham de Moivre.[282]

For random walks in  -dimensional integer lattices, George Pólya published, in 1919 and 1921, work where he studied the probability of a symmetric random walk returning to a previous position in the lattice. Pólya showed that a symmetric random walk, which has an equal probability to advance in any direction in the lattice, will return to a previous position in the lattice an infinite number of times with probability one in one and two dimensions, but with probability zero in three or higher dimensions.[283][284]

Wiener process edit

The Wiener process or Brownian motion process has its origins in different fields including statistics, finance and physics.[21] In 1880, Danish astronomer Thorvald Thiele wrote a paper on the method of least squares, where he used the process to study the errors of a model in time-series analysis.[285][286][287] The work is now considered as an early discovery of the statistical method known as Kalman filtering, but the work was largely overlooked. It is thought that the ideas in Thiele's paper were too advanced to have been understood by the broader mathematical and statistical community at the time.[287]

 
Norbert Wiener gave the first mathematical proof of the existence of the Wiener process. This mathematical object had appeared previously in the work of Thorvald Thiele, Louis Bachelier, and Albert Einstein.[21]

The French mathematician Louis Bachelier used a Wiener process in his 1900 thesis[288][289] in order to model price changes on the Paris Bourse, a stock exchange,[290] without knowing the work of Thiele.[21] It has been speculated that Bachelier drew ideas from the random walk model of Jules Regnault, but Bachelier did not cite him,[291] and Bachelier's thesis is now considered pioneering in the field of financial mathematics.[290][291]

It is commonly thought that Bachelier's work gained little attention and was forgotten for decades until it was rediscovered in the 1950s by the Leonard Savage, and then become more popular after Bachelier's thesis was translated into English in 1964. But the work was never forgotten in the mathematical community, as Bachelier published a book in 1912 detailing his ideas,[291] which was cited by mathematicians including Doob, Feller[291] and Kolmogorov.[21] The book continued to be cited, but then starting in the 1960s, the original thesis by Bachelier began to be cited more than his book when economists started citing Bachelier's work.[291]

In 1905, Albert Einstein published a paper where he studied the physical observation of Brownian motion or movement to explain the seemingly random movements of particles in liquids by using ideas from the kinetic theory of gases. Einstein derived a differential equation, known as a diffusion equation, for describing the probability of finding a particle in a certain region of space. Shortly after Einstein's first paper on Brownian movement, Marian Smoluchowski published work where he cited Einstein, but wrote that he had independently derived the equivalent results by using a different method.[292]

Einstein's work, as well as experimental results obtained by Jean Perrin, later inspired Norbert Wiener in the 1920s[293] to use a type of measure theory, developed by Percy Daniell, and Fourier analysis to prove the existence of the Wiener process as a mathematical object.[21]

Poisson process edit

The Poisson process is named after Siméon Poisson, due to its definition involving the Poisson distribution, but Poisson never studied the process.[22][294] There are a number of claims for early uses or discoveries of the Poisson process.[22][24] At the beginning of the 20th century, the Poisson process would arise independently in different situations.[22][24] In Sweden 1903, Filip Lundberg published a thesis containing work, now considered fundamental and pioneering, where he proposed to model insurance claims with a homogeneous Poisson process.[295][296]

Another discovery occurred in Denmark in 1909 when A.K. Erlang derived the Poisson distribution when developing a mathematical model for the number of incoming phone calls in a finite time interval. Erlang was not at the time aware of Poisson's earlier work and assumed that the number phone calls arriving in each interval of time were independent to each other. He then found the limiting case, which is effectively recasting the Poisson distribution as a limit of the binomial distribution.[22]

In 1910, Ernest Rutherford and Hans Geiger published experimental results on counting alpha particles. Motivated by their work, Harry Bateman studied the counting problem and derived Poisson probabilities as a solution to a family of differential equations, resulting in the independent discovery of the Poisson process.[22] After this time there were many studies and applications of the Poisson process, but its early history is complicated, which has been explained by the various applications of the process in numerous fields by biologists, ecologists, engineers and various physical scientists.[22]

Markov processes edit

Markov processes and Markov chains are named after Andrey Markov who studied Markov chains in the early 20th century. Markov was interested in studying an extension of independent random sequences. In his first paper on Markov chains, published in 1906, Markov showed that under certain conditions the average outcomes of the Markov chain would converge to a fixed vector of values, so proving a weak law of large numbers without the independence assumption,[297][298][299] which had been commonly regarded as a requirement for such mathematical laws to hold.[299] Markov later used Markov chains to study the distribution of vowels in Eugene Onegin, written by Alexander Pushkin, and proved a central limit theorem for such chains.[300][297]

In 1912, Poincaré studied Markov chains on finite groups with an aim to study card shuffling. Other early uses of Markov chains include a diffusion model, introduced by Paul and Tatyana Ehrenfest in 1907, and a branching process, introduced by Francis Galton and Henry William Watson in 1873, preceding the work of Markov.[297][298] After the work of Galton and Watson, it was later revealed that their branching process had been independently discovered and studied around three decades earlier by Irénée-Jules Bienaymé.[301] Starting in 1928, Maurice Fréchet became interested in Markov chains, eventually resulting in him publishing in 1938 a detailed study on Markov chains.[297][302]

Andrei Kolmogorov developed in a 1931 paper a large part of the early theory of continuous-time Markov processes.[252][258] Kolmogorov was partly inspired by Louis Bachelier's 1900 work on fluctuations in the stock market as well as Norbert Wiener's work on Einstein's model of Brownian movement.[258][303] He introduced and studied a particular set of Markov processes known as diffusion processes, where he derived a set of differential equations describing the processes.[258][304] Independent of Kolmogorov's work, Sydney Chapman derived in a 1928 paper an equation, now called the Chapman–Kolmogorov equation, in a less mathematically rigorous way than Kolmogorov, while studying Brownian movement.[305] The differential equations are now called the Kolmogorov equations[306] or the Kolmogorov–Chapman equations.[307] Other mathematicians who contributed significantly to the foundations of Markov processes include William Feller, starting in the 1930s, and then later Eugene Dynkin, starting in the 1950s.[252]

Lévy processes edit

Lévy processes such as the Wiener process and the Poisson process (on the real line) are named after Paul Lévy who started studying them in the 1930s,[225] but they have connections to infinitely divisible distributions going back to the 1920s.[224] In a 1932 paper, Kolmogorov derived a characteristic function for random variables associated with Lévy processes. This result was later derived under more general conditions by Lévy in 1934, and then Khinchin independently gave an alternative form for this characteristic function in 1937.[252][308] In addition to Lévy, Khinchin and Kolomogrov, early fundamental contributions to the theory of Lévy processes were made by Bruno de Finetti and Kiyosi Itô.[224]

Mathematical construction edit

In mathematics, constructions of mathematical objects are needed, which is also the case for stochastic processes, to prove that they exist mathematically.[57] There are two main approaches for constructing a stochastic process. One approach involves considering a measurable space of functions, defining a suitable measurable mapping from a probability space to this measurable space of functions, and then deriving the corresponding finite-dimensional distributions.[309]

Another approach involves defining a collection of random variables to have specific finite-dimensional distributions, and then using Kolmogorov's existence theorem[j] to prove a corresponding stochastic process exists.[57][309] This theorem, which is an existence theorem for measures on infinite product spaces,[313] says that if any finite-dimensional distributions satisfy two conditions, known as consistency conditions, then there exists a stochastic process with those finite-dimensional distributions.[57]

Construction issues edit

When constructing continuous-time stochastic processes certain mathematical difficulties arise, due to the uncountable index sets, which do not occur with discrete-time processes.[58][59] One problem is that is it possible to have more than one stochastic process with the same finite-dimensional distributions. For example, both the left-continuous modification and the right-continuous modification of a Poisson process have the same finite-dimensional distributions.[314] This means that the distribution of the stochastic process does not, necessarily, specify uniquely the properties of the sample functions of the stochastic process.[309][315]

Another problem is that functionals of continuous-time process that rely upon an uncountable number of points of the index set may not be measurable, so the probabilities of certain events may not be well-defined.[168] For example, the supremum of a stochastic process or random field is not necessarily a well-defined random variable.[30][59] For a continuous-time stochastic process  , other characteristics that depend on an uncountable number of points of the index set   include:[168]

  • a sample function of a stochastic process   is a continuous function of  ;
  • a sample function of a stochastic process   is a bounded function of  ; and
  • a sample function of a stochastic process   is an increasing function of  .

To overcome these two difficulties, different assumptions and approaches are possible.[69]

Resolving construction issues edit

One approach for avoiding mathematical construction issues of stochastic processes, proposed by Joseph Doob, is to assume that the stochastic process is separable.[316] Separability ensures that infinite-dimensional distributions determine the properties of sample functions by requiring that sample functions are essentially determined by their values on a dense countable set of points in the index set.[317] Furthermore, if a stochastic process is separable, then functionals of an uncountable number of points of the index set are measurable and their probabilities can be studied.[168][317]

Another approach is possible, originally developed by Anatoliy Skorokhod and Andrei Kolmogorov,[318] for a continuous-time stochastic process with any metric space as its state space. For the construction of such a stochastic process, it is assumed that the sample functions of the stochastic process belong to some suitable function space, which is usually the Skorokhod space consisting of all right-continuous functions with left limits. This approach is now more used than the separability assumption,[69][263] but such a stochastic process based on this approach will be automatically separable.[319]

Although less used, the separability assumption is considered more general because every stochastic process has a separable version.[263] It is also used when it is not possible to construct a stochastic process in a Skorokhod space.[173] For example, separability is assumed when constructing and studying random fields, where the collection of random variables is now indexed by sets other than the real line such as  -dimensional Euclidean space.[30][320]

See also edit

Notes edit

  1. ^ The term Brownian motion can refer to the physical process, also known as Brownian movement, and the stochastic process, a mathematical object, but to avoid ambiguity this article uses the terms Brownian motion process or Wiener process for the latter in a style similar to, for example, Gikhman and Skorokhod[19] or Rosenblatt.[20]
  2. ^ The term "separable" appears twice here with two different meanings, where the first meaning is from probability and the second from topology and analysis. For a stochastic process to be separable (in a probabilistic sense), its index set must be a separable space (in a topological or analytic sense), in addition to other conditions.[136]
  3. ^ The definition of separability for a continuous-time real-valued stochastic process can be stated in other ways.[172][173]
  4. ^ In the context of point processes, the term "state space" can mean the space on which the point process is defined such as the real line,[234][235] which corresponds to the index set in stochastic process terminology.
  5. ^ Also known as James or Jacques Bernoulli.[245]
  6. ^ It has been remarked that a notable exception was the St Petersburg School in Russia, where mathematicians led by Chebyshev studied probability theory.[250]
  7. ^ The name Khinchin is also written in (or transliterated into) English as Khintchine.[63]
  8. ^ Doob, when citing Khinchin, uses the term 'chance variable', which used to be an alternative term for 'random variable'.[261]
  9. ^ Later translated into English and published in 1950 as Foundations of the Theory of Probability[249]
  10. ^ The theorem has other names including Kolmogorov's consistency theorem,[310] Kolmogorov's extension theorem[311] or the Daniell–Kolmogorov theorem.[312]

References edit

  1. ^ a b c d e f g h i Joseph L. Doob (1990). Stochastic processes. Wiley. pp. 46, 47.
  2. ^ a b c d L. C. G. Rogers; David Williams (2000). Diffusions, Markov Processes, and Martingales: Volume 1, Foundations. Cambridge University Press. p. 1. ISBN 978-1-107-71749-7.
  3. ^ a b c J. Michael Steele (2012). Stochastic Calculus and Financial Applications. Springer Science & Business Media. p. 29. ISBN 978-1-4684-9305-4.
  4. ^ a b c d e Emanuel Parzen (2015). Stochastic Processes. Courier Dover Publications. pp. 7, 8. ISBN 978-0-486-79688-8.
  5. ^ a b c d e f g h i j k l Iosif Ilyich Gikhman; Anatoly Vladimirovich Skorokhod (1969). Introduction to the Theory of Random Processes. Courier Corporation. p. 1. ISBN 978-0-486-69387-3.
  6. ^ Bressloff, Paul C. (2014). Stochastic Processes in Cell Biology. Springer. ISBN 978-3-319-08488-6.
  7. ^ Van Kampen, N. G. (2011). Stochastic Processes in Physics and Chemistry. Elsevier. ISBN 978-0-08-047536-3.
  8. ^ Lande, Russell; Engen, Steinar; Sæther, Bernt-Erik (2003). Stochastic Population Dynamics in Ecology and Conservation. Oxford University Press. ISBN 978-0-19-852525-7.
  9. ^ Laing, Carlo; Lord, Gabriel J. (2010). Stochastic Methods in Neuroscience. Oxford University Press. ISBN 978-0-19-923507-0.
  10. ^ Paul, Wolfgang; Baschnagel, Jörg (2013). Stochastic Processes: From Physics to Finance. Springer Science+Business Media. ISBN 978-3-319-00327-6.
  11. ^ Dougherty, Edward R. (1999). Random processes for image and signal processing. SPIE Optical Engineering Press. ISBN 978-0-8194-2513-3.
  12. ^ Bertsekas, Dimitri P. (1996). Stochastic Optimal Control: The Discrete-Time Case. Athena Scientific. ISBN 1-886529-03-5.
  13. ^ Thomas M. Cover; Joy A. Thomas (2012). Elements of Information Theory. John Wiley & Sons. p. 71. ISBN 978-1-118-58577-1.
  14. ^ Baron, Michael (2015). Probability and Statistics for Computer Scientists (2nd ed.). CRC Press. p. 131. ISBN 978-1-4987-6060-7.
  15. ^ Baccelli, François; Blaszczyszyn, Bartlomiej (2009). Stochastic Geometry and Wireless Networks. Now Publishers Inc. ISBN 978-1-60198-264-3.
  16. ^ Steele, J. Michael (2001). Stochastic Calculus and Financial Applications. Springer Science+Business Media. ISBN 978-0-387-95016-7.
  17. ^ a b Musiela, Marek; Rutkowski, Marek (2006). Martingale Methods in Financial Modelling. Springer Science+Business Media. ISBN 978-3-540-26653-2.
  18. ^ Shreve, Steven E. (2004). Stochastic Calculus for Finance II: Continuous-Time Models. Springer Science+Business Media. ISBN 978-0-387-40101-0.
  19. ^ Iosif Ilyich Gikhman; Anatoly Vladimirovich Skorokhod (1969). Introduction to the Theory of Random Processes. Courier Corporation. ISBN 978-0-486-69387-3.
  20. ^ Murray Rosenblatt (1962). Random Processes. Oxford University Press.
  21. ^ a b c d e f g h i Jarrow, Robert; Protter, Philip (2004). "A short history of stochastic integration and mathematical finance: the early years, 1880–1970". A Festschrift for Herman Rubin. Institute of Mathematical Statistics Lecture Notes - Monograph Series. pp. 75–80. CiteSeerX 10.1.1.114.632. doi:10.1214/lnms/1196285381. ISBN 978-0-940600-61-4. ISSN 0749-2170.
  22. ^ a b c d e f g h Stirzaker, David (2000). "Advice to Hedgehogs, or, Constants Can Vary". The Mathematical Gazette. 84 (500): 197–210. doi:10.2307/3621649. ISSN 0025-5572. JSTOR 3621649. S2CID 125163415.
  23. ^ Donald L. Snyder; Michael I. Miller (2012). Random Point Processes in Time and Space. Springer Science & Business Media. p. 32. ISBN 978-1-4612-3166-0.
  24. ^ a b c d Guttorp, Peter; Thorarinsdottir, Thordis L. (2012). "What Happened to Discrete Chaos, the Quenouille Process, and the Sharp Markov Property? Some History of Stochastic Point Processes". International Statistical Review. 80 (2): 253–268. doi:10.1111/j.1751-5823.2012.00181.x. ISSN 0306-7734. S2CID 80836.
  25. ^ Gusak, Dmytro; Kukush, Alexander; Kulik, Alexey; Mishura, Yuliya; Pilipenko, Andrey (2010). Theory of Stochastic Processes: With Applications to Financial Mathematics and Risk Theory. Springer Science & Business Media. p. 21. ISBN 978-0-387-87862-1.
  26. ^ Valeriy Skorokhod (2005). Basic Principles and Applications of Probability Theory. Springer Science & Business Media. p. 42. ISBN 978-3-540-26312-8.
  27. ^ a b c d e f Olav Kallenberg (2002). Foundations of Modern Probability. Springer Science & Business Media. pp. 24–25. ISBN 978-0-387-95313-7.
  28. ^ a b c d e f g h i j k l m n o p John Lamperti (1977). Stochastic processes: a survey of the mathematical theory. Springer-Verlag. pp. 1–2. ISBN 978-3-540-90275-1.
  29. ^ a b c d Loïc Chaumont; Marc Yor (2012). Exercises in Probability: A Guided Tour from Measure Theory to Random Processes, Via Conditioning. Cambridge University Press. p. 175. ISBN 978-1-107-60655-5.
  30. ^ a b c d e f g h Robert J. Adler; Jonathan E. Taylor (2009). Random Fields and Geometry. Springer Science & Business Media. pp. 7–8. ISBN 978-0-387-48116-6.
  31. ^ Gregory F. Lawler; Vlada Limic (2010). Random Walk: A Modern Introduction. Cambridge University Press. ISBN 978-1-139-48876-1.
  32. ^ David Williams (1991). Probability with Martingales. Cambridge University Press. ISBN 978-0-521-40605-5.
  33. ^ L. C. G. Rogers; David Williams (2000). Diffusions, Markov Processes, and Martingales: Volume 1, Foundations. Cambridge University Press. ISBN 978-1-107-71749-7.
  34. ^ David Applebaum (2004). Lévy Processes and Stochastic Calculus. Cambridge University Press. ISBN 978-0-521-83263-2.
  35. ^ Mikhail Lifshits (2012). Lectures on Gaussian Processes. Springer Science & Business Media. ISBN 978-3-642-24939-6.
  36. ^ Robert J. Adler (2010). The Geometry of Random Fields. SIAM. ISBN 978-0-89871-693-1.
  37. ^ Samuel Karlin; Howard E. Taylor (2012). A First Course in Stochastic Processes. Academic Press. ISBN 978-0-08-057041-9.
  38. ^ Bruce Hajek (2015). Random Processes for Engineers. Cambridge University Press. ISBN 978-1-316-24124-0.
  39. ^ a b G. Latouche; V. Ramaswami (1999). Introduction to Matrix Analytic Methods in Stochastic Modeling. SIAM. ISBN 978-0-89871-425-8.
  40. ^ D.J. Daley; David Vere-Jones (2007). An Introduction to the Theory of Point Processes: Volume II: General Theory and Structure. Springer Science & Business Media. ISBN 978-0-387-21337-8.
  41. ^ Patrick Billingsley (2008). Probability and Measure. Wiley India Pvt. Limited. ISBN 978-81-265-1771-8.
  42. ^ Pierre Brémaud (2014). Fourier Analysis and Stochastic Processes. Springer. ISBN 978-3-319-09590-5.
  43. ^ Adam Bobrowski (2005). Functional Analysis for Probability and Stochastic Processes: An Introduction. Cambridge University Press. ISBN 978-0-521-83166-6.
  44. ^ Applebaum, David (2004). "Lévy processes: From probability to finance and quantum groups". Notices of the AMS. 51 (11): 1336–1347.
  45. ^ a b Jochen Blath; Peter Imkeller; Sylvie Roelly (2011). Surveys in Stochastic Processes. European Mathematical Society. ISBN 978-3-03719-072-2.
  46. ^ Michel Talagrand (2014). Upper and Lower Bounds for Stochastic Processes: Modern Methods and Classical Problems. Springer Science & Business Media. pp. 4–. ISBN 978-3-642-54075-2.
  47. ^ Paul C. Bressloff (2014). Stochastic Processes in Cell Biology. Springer. pp. vii–ix. ISBN 978-3-319-08488-6.
  48. ^ a b c d Samuel Karlin; Howard E. Taylor (2012). A First Course in Stochastic Processes. Academic Press. p. 27. ISBN 978-0-08-057041-9.
  49. ^ a b c d e f g h i j Applebaum, David (2004). "Lévy processes: From probability to finance and quantum groups". Notices of the AMS. 51 (11): 1337.
  50. ^ a b L. C. G. Rogers; David Williams (2000). Diffusions, Markov Processes, and Martingales: Volume 1, Foundations. Cambridge University Press. pp. 121–124. ISBN 978-1-107-71749-7.
  51. ^ a b c d e f Ionut Florescu (2014). Probability and Stochastic Processes. John Wiley & Sons. pp. 294, 295. ISBN 978-1-118-59320-2.
  52. ^ a b Samuel Karlin; Howard E. Taylor (2012). A First Course in Stochastic Processes. Academic Press. p. 26. ISBN 978-0-08-057041-9.
  53. ^ Donald L. Snyder; Michael I. Miller (2012). Random Point Processes in Time and Space. Springer Science & Business Media. pp. 24, 25. ISBN 978-1-4612-3166-0.
  54. ^ a b Patrick Billingsley (2008). Probability and Measure. Wiley India Pvt. Limited. p. 482. ISBN 978-81-265-1771-8.
  55. ^ a b Alexander A. Borovkov (2013). Probability Theory. Springer Science & Business Media. p. 527. ISBN 978-1-4471-5201-9.
  56. ^ a b c Pierre Brémaud (2014). Fourier Analysis and Stochastic Processes. Springer. p. 120. ISBN 978-3-319-09590-5.
  57. ^ a b c d e Jeffrey S Rosenthal (2006). A First Look at Rigorous Probability Theory. World Scientific Publishing Co Inc. pp. 177–178. ISBN 978-981-310-165-4.
  58. ^ a b Peter E. Kloeden; Eckhard Platen (2013). Numerical Solution of Stochastic Differential Equations. Springer Science & Business Media. p. 63. ISBN 978-3-662-12616-5.
  59. ^ a b c Davar Khoshnevisan (2006). Multiparameter Processes: An Introduction to Random Fields. Springer Science & Business Media. pp. 153–155. ISBN 978-0-387-21631-7.
  60. ^ a b "Stochastic". Oxford English Dictionary (Online ed.). Oxford University Press. (Subscription or participating institution membership required.)
  61. ^ O. B. Sheĭnin (2006). Theory of probability and statistics as exemplified in short dictums. NG Verlag. p. 5. ISBN 978-3-938417-40-9.
  62. ^ Oscar Sheynin; Heinrich Strecker (2011). Alexandr A. Chuprov: Life, Work, Correspondence. V&R unipress GmbH. p. 136. ISBN 978-3-89971-812-6.
  63. ^ a b c d Doob, Joseph (1934). "Stochastic Processes and Statistics". Proceedings of the National Academy of Sciences of the United States of America. 20 (6): 376–379. Bibcode:1934PNAS...20..376D. doi:10.1073/pnas.20.6.376. PMC 1076423. PMID 16587907.
  64. ^ Khintchine, A. (1934). "Korrelationstheorie der stationeren stochastischen Prozesse". Mathematische Annalen. 109 (1): 604–615. doi:10.1007/BF01449156. ISSN 0025-5831. S2CID 122842868.
stochastic, process, probability, theory, related, fields, stochastic, random, process, mathematical, object, usually, defined, sequence, random, variables, probability, space, where, index, sequence, often, interpretation, time, widely, used, mathematical, mo. In probability theory and related fields a stochastic s t e ˈ k ae s t ɪ k or random process is a mathematical object usually defined as a sequence of random variables in a probability space where the index of the sequence often has the interpretation of time Stochastic processes are widely used as mathematical models of systems and phenomena that appear to vary in a random manner Examples include the growth of a bacterial population an electrical current fluctuating due to thermal noise or the movement of a gas molecule 1 4 5 Stochastic processes have applications in many disciplines such as biology 6 chemistry 7 ecology 8 neuroscience 9 physics 10 image processing signal processing 11 control theory 12 information theory 13 computer science 14 and telecommunications 15 Furthermore seemingly random changes in financial markets have motivated the extensive use of stochastic processes in finance 16 17 18 A computer simulated realization of a Wiener or Brownian motion process on the surface of a sphere The Wiener process is widely considered the most studied and central stochastic process in probability theory 1 2 3 Applications and the study of phenomena have in turn inspired the proposal of new stochastic processes Examples of such stochastic processes include the Wiener process or Brownian motion process a used by Louis Bachelier to study price changes on the Paris Bourse 21 and the Poisson process used by A K Erlang to study the number of phone calls occurring in a certain period of time 22 These two stochastic processes are considered the most important and central in the theory of stochastic processes 1 4 23 and were invented repeatedly and independently both before and after Bachelier and Erlang in different settings and countries 21 24 The term random function is also used to refer to a stochastic or random process 25 26 because a stochastic process can also be interpreted as a random element in a function space 27 28 The terms stochastic process and random process are used interchangeably often with no specific mathematical space for the set that indexes the random variables 27 29 But often these two terms are used when the random variables are indexed by the integers or an interval of the real line 5 29 If the random variables are indexed by the Cartesian plane or some higher dimensional Euclidean space then the collection of random variables is usually called a random field instead 5 30 The values of a stochastic process are not always numbers and can be vectors or other mathematical objects 5 28 Based on their mathematical properties stochastic processes can be grouped into various categories which include random walks 31 martingales 32 Markov processes 33 Levy processes 34 Gaussian processes 35 random fields 36 renewal processes and branching processes 37 The study of stochastic processes uses mathematical knowledge and techniques from probability calculus linear algebra set theory and topology 38 39 40 as well as branches of mathematical analysis such as real analysis measure theory Fourier analysis and functional analysis 41 42 43 The theory of stochastic processes is considered to be an important contribution to mathematics 44 and it continues to be an active topic of research for both theoretical reasons and applications 45 46 47 Contents 1 Introduction 1 1 Classifications 1 2 Etymology 1 3 Terminology 1 4 Notation 2 Examples 2 1 Bernoulli process 2 2 Random walk 2 3 Wiener process 2 4 Poisson process 3 Definitions 3 1 Stochastic process 3 2 Index set 3 3 State space 3 4 Sample function 3 5 Increment 3 6 Further definitions 3 6 1 Law 3 6 2 Finite dimensional probability distributions 3 6 3 Stationarity 3 6 4 Filtration 3 6 5 Modification 3 6 6 Indistinguishable 3 6 7 Separability 3 6 8 Independence 3 6 9 Uncorrelatedness 3 6 10 Independence implies uncorrelatedness 3 6 11 Orthogonality 3 6 12 Skorokhod space 3 6 13 Regularity 4 Further examples 4 1 Markov processes and chains 4 2 Martingale 4 3 Levy process 4 4 Random field 4 5 Point process 5 History 5 1 Early probability theory 5 2 Statistical mechanics 5 3 Measure theory and probability theory 5 4 Birth of modern probability theory 5 5 Stochastic processes after World War II 5 6 Discoveries of specific stochastic processes 5 6 1 Bernoulli process 5 6 2 Random walks 5 6 3 Wiener process 5 6 4 Poisson process 5 6 5 Markov processes 5 6 6 Levy processes 6 Mathematical construction 6 1 Construction issues 6 2 Resolving construction issues 7 See also 8 Notes 9 References 10 Further reading 10 1 Articles 10 2 Books 11 External linksIntroduction editA stochastic or random process can be defined as a collection of random variables that is indexed by some mathematical set meaning that each random variable of the stochastic process is uniquely associated with an element in the set 4 5 The set used to index the random variables is called the index set Historically the index set was some subset of the real line such as the natural numbers giving the index set the interpretation of time 1 Each random variable in the collection takes values from the same mathematical space known as the state space This state space can be for example the integers the real line or n displaystyle n nbsp dimensional Euclidean space 1 5 An increment is the amount that a stochastic process changes between two index values often interpreted as two points in time 48 49 A stochastic process can have many outcomes due to its randomness and a single outcome of a stochastic process is called among other names a sample function or realization 28 50 nbsp A single computer simulated sample function or realization among other terms of a three dimensional Wiener or Brownian motion process for time 0 t 2 The index set of this stochastic process is the non negative numbers while its state space is three dimensional Euclidean space Classifications edit A stochastic process can be classified in different ways for example by its state space its index set or the dependence among the random variables One common way of classification is by the cardinality of the index set and the state space 51 52 53 When interpreted as time if the index set of a stochastic process has a finite or countable number of elements such as a finite set of numbers the set of integers or the natural numbers then the stochastic process is said to be in discrete time 54 55 If the index set is some interval of the real line then time is said to be continuous The two types of stochastic processes are respectively referred to as discrete time and continuous time stochastic processes 48 56 57 Discrete time stochastic processes are considered easier to study because continuous time processes require more advanced mathematical techniques and knowledge particularly due to the index set being uncountable 58 59 If the index set is the integers or some subset of them then the stochastic process can also be called a random sequence 55 If the state space is the integers or natural numbers then the stochastic process is called a discrete or integer valued stochastic process If the state space is the real line then the stochastic process is referred to as a real valued stochastic process or a process with continuous state space If the state space is n displaystyle n nbsp dimensional Euclidean space then the stochastic process is called a n displaystyle n nbsp dimensional vector process or n displaystyle n nbsp vector process 51 52 Etymology edit The word stochastic in English was originally used as an adjective with the definition pertaining to conjecturing and stemming from a Greek word meaning to aim at a mark guess and the Oxford English Dictionary gives the year 1662 as its earliest occurrence 60 In his work on probability Ars Conjectandi originally published in Latin in 1713 Jakob Bernoulli used the phrase Ars Conjectandi sive Stochastice which has been translated to the art of conjecturing or stochastics 61 This phrase was used with reference to Bernoulli by Ladislaus Bortkiewicz 62 who in 1917 wrote in German the word stochastik with a sense meaning random The term stochastic process first appeared in English in a 1934 paper by Joseph Doob 60 For the term and a specific mathematical definition Doob cited another 1934 paper where the term stochastischer Prozess was used in German by Aleksandr Khinchin 63 64 though the German term had been used earlier for example by Andrei Kolmogorov in 1931 65 According to the Oxford English Dictionary early occurrences of the word random in English with its current meaning which relates to chance or luck date back to the 16th century while earlier recorded usages started in the 14th century as a noun meaning impetuosity great speed force or violence in riding running striking etc The word itself comes from a Middle French word meaning speed haste and it is probably derived from a French verb meaning to run or to gallop The first written appearance of the term random process pre dates stochastic process which the Oxford English Dictionary also gives as a synonym and was used in an article by Francis Edgeworth published in 1888 66 Terminology edit The definition of a stochastic process varies 67 but a stochastic process is traditionally defined as a collection of random variables indexed by some set 68 69 The terms random process and stochastic process are considered synonyms and are used interchangeably without the index set being precisely specified 27 29 30 70 71 72 Both collection 28 70 or family are used 4 73 while instead of index set sometimes the terms parameter set 28 or parameter space 30 are used The term random function is also used to refer to a stochastic or random process 5 74 75 though sometimes it is only used when the stochastic process takes real values 28 73 This term is also used when the index sets are mathematical spaces other than the real line 5 76 while the terms stochastic process and random process are usually used when the index set is interpreted as time 5 76 77 and other terms are used such as random field when the index set is n displaystyle n nbsp dimensional Euclidean space R n displaystyle mathbb R n nbsp or a manifold 5 28 30 Notation edit A stochastic process can be denoted among other ways by X t t T displaystyle X t t in T nbsp 56 X t t T displaystyle X t t in T nbsp 69 X t displaystyle X t nbsp 78 X t displaystyle X t nbsp or simply as X displaystyle X nbsp Some authors mistakenly write X t displaystyle X t nbsp even though it is an abuse of function notation 79 For example X t displaystyle X t nbsp or X t displaystyle X t nbsp are used to refer to the random variable with the index t displaystyle t nbsp and not the entire stochastic process 78 If the index set is T 0 displaystyle T 0 infty nbsp then one can write for example X t t 0 displaystyle X t t geq 0 nbsp to denote the stochastic process 29 Examples editBernoulli process edit Main article Bernoulli process One of the simplest stochastic processes is the Bernoulli process 80 which is a sequence of independent and identically distributed iid random variables where each random variable takes either the value one or zero say one with probability p displaystyle p nbsp and zero with probability 1 p displaystyle 1 p nbsp This process can be linked to an idealisation of repeatedly flipping a coin where the probability of obtaining a head is taken to be p displaystyle p nbsp and its value is one while the value of a tail is zero 81 In other words a Bernoulli process is a sequence of iid Bernoulli random variables 82 where each idealised coin flip is an example of a Bernoulli trial 83 Random walk edit Main article Random walk Random walks are stochastic processes that are usually defined as sums of iid random variables or random vectors in Euclidean space so they are processes that change in discrete time 84 85 86 87 88 But some also use the term to refer to processes that change in continuous time 89 particularly the Wiener process used in financial models which has led to some confusion resulting in its criticism 90 There are other various types of random walks defined so their state spaces can be other mathematical objects such as lattices and groups and in general they are highly studied and have many applications in different disciplines 89 91 A classic example of a random walk is known as the simple random walk which is a stochastic process in discrete time with the integers as the state space and is based on a Bernoulli process where each Bernoulli variable takes either the value positive one or negative one In other words the simple random walk takes place on the integers and its value increases by one with probability say p displaystyle p nbsp or decreases by one with probability 1 p displaystyle 1 p nbsp so the index set of this random walk is the natural numbers while its state space is the integers If p 0 5 displaystyle p 0 5 nbsp this random walk is called a symmetric random walk 92 93 Wiener process edit Main article Wiener process The Wiener process is a stochastic process with stationary and independent increments that are normally distributed based on the size of the increments 2 94 The Wiener process is named after Norbert Wiener who proved its mathematical existence but the process is also called the Brownian motion process or just Brownian motion due to its historical connection as a model for Brownian movement in liquids 95 96 97 nbsp Realizations of Wiener processes or Brownian motion processes with drift blue and without drift red Playing a central role in the theory of probability the Wiener process is often considered the most important and studied stochastic process with connections to other stochastic processes 1 2 3 98 99 100 101 Its index set and state space are the non negative numbers and real numbers respectively so it has both continuous index set and states space 102 But the process can be defined more generally so its state space can be n displaystyle n nbsp dimensional Euclidean space 91 99 103 If the mean of any increment is zero then the resulting Wiener or Brownian motion process is said to have zero drift If the mean of the increment for any two points in time is equal to the time difference multiplied by some constant m displaystyle mu nbsp which is a real number then the resulting stochastic process is said to have drift m displaystyle mu nbsp 104 105 106 Almost surely a sample path of a Wiener process is continuous everywhere but nowhere differentiable It can be considered as a continuous version of the simple random walk 49 105 The process arises as the mathematical limit of other stochastic processes such as certain random walks rescaled 107 108 which is the subject of Donsker s theorem or invariance principle also known as the functional central limit theorem 109 110 111 The Wiener process is a member of some important families of stochastic processes including Markov processes Levy processes and Gaussian processes 2 49 The process also has many applications and is the main stochastic process used in stochastic calculus 112 113 It plays a central role in quantitative finance 114 115 where it is used for example in the Black Scholes Merton model 116 The process is also used in different fields including the majority of natural sciences as well as some branches of social sciences as a mathematical model for various random phenomena 3 117 118 Poisson process edit Main article Poisson process The Poisson process is a stochastic process that has different forms and definitions 119 120 It can be defined as a counting process which is a stochastic process that represents the random number of points or events up to some time The number of points of the process that are located in the interval from zero to some given time is a Poisson random variable that depends on that time and some parameter This process has the natural numbers as its state space and the non negative numbers as its index set This process is also called the Poisson counting process since it can be interpreted as an example of a counting process 119 If a Poisson process is defined with a single positive constant then the process is called a homogeneous Poisson process 119 121 The homogeneous Poisson process is a member of important classes of stochastic processes such as Markov processes and Levy processes 49 The homogeneous Poisson process can be defined and generalized in different ways It can be defined such that its index set is the real line and this stochastic process is also called the stationary Poisson process 122 123 If the parameter constant of the Poisson process is replaced with some non negative integrable function of t displaystyle t nbsp the resulting process is called an inhomogeneous or nonhomogeneous Poisson process where the average density of points of the process is no longer constant 124 Serving as a fundamental process in queueing theory the Poisson process is an important process for mathematical models where it finds applications for models of events randomly occurring in certain time windows 125 126 Defined on the real line the Poisson process can be interpreted as a stochastic process 49 127 among other random objects 128 129 But then it can be defined on the n displaystyle n nbsp dimensional Euclidean space or other mathematical spaces 130 where it is often interpreted as a random set or a random counting measure instead of a stochastic process 128 129 In this setting the Poisson process also called the Poisson point process is one of the most important objects in probability theory both for applications and theoretical reasons 22 131 But it has been remarked that the Poisson process does not receive as much attention as it should partly due to it often being considered just on the real line and not on other mathematical spaces 131 132 Definitions editStochastic process edit A stochastic process is defined as a collection of random variables defined on a common probability space W F P displaystyle Omega mathcal F P nbsp where W displaystyle Omega nbsp is a sample space F displaystyle mathcal F nbsp is a s displaystyle sigma nbsp algebra and P displaystyle P nbsp is a probability measure and the random variables indexed by some set T displaystyle T nbsp all take values in the same mathematical space S displaystyle S nbsp which must be measurable with respect to some s displaystyle sigma nbsp algebra S displaystyle Sigma nbsp 28 In other words for a given probability space W F P displaystyle Omega mathcal F P nbsp and a measurable space S S displaystyle S Sigma nbsp a stochastic process is a collection of S displaystyle S nbsp valued random variables which can be written as 80 X t t T displaystyle X t t in T nbsp Historically in many problems from the natural sciences a point t T displaystyle t in T nbsp had the meaning of time so X t displaystyle X t nbsp is a random variable representing a value observed at time t displaystyle t nbsp 133 A stochastic process can also be written as X t w t T displaystyle X t omega t in T nbsp to reflect that it is actually a function of two variables t T displaystyle t in T nbsp and w W displaystyle omega in Omega nbsp 28 134 There are other ways to consider a stochastic process with the above definition being considered the traditional one 68 69 For example a stochastic process can be interpreted or defined as a S T displaystyle S T nbsp valued random variable where S T displaystyle S T nbsp is the space of all the possible functions from the set T displaystyle T nbsp into the space S displaystyle S nbsp 27 68 However this alternative definition as a function valued random variable in general requires additional regularity assumptions to be well defined 135 Index set edit The set T displaystyle T nbsp is called the index set 4 51 or parameter set 28 136 of the stochastic process Often this set is some subset of the real line such as the natural numbers or an interval giving the set T displaystyle T nbsp the interpretation of time 1 In addition to these sets the index set T displaystyle T nbsp can be another set with a total order or a more general set 1 54 such as the Cartesian plane R 2 displaystyle mathbb R 2 nbsp or n displaystyle n nbsp dimensional Euclidean space where an element t T displaystyle t in T nbsp can represent a point in space 48 137 That said many results and theorems are only possible for stochastic processes with a totally ordered index set 138 State space edit The mathematical space S displaystyle S nbsp of a stochastic process is called its state space This mathematical space can be defined using integers real lines n displaystyle n nbsp dimensional Euclidean spaces complex planes or more abstract mathematical spaces The state space is defined using elements that reflect the different values that the stochastic process can take 1 5 28 51 56 Sample function edit A sample function is a single outcome of a stochastic process so it is formed by taking a single possible value of each random variable of the stochastic process 28 139 More precisely if X t w t T displaystyle X t omega t in T nbsp is a stochastic process then for any point w W displaystyle omega in Omega nbsp the mapping X w T S displaystyle X cdot omega T rightarrow S nbsp is called a sample function a realization or particularly when T displaystyle T nbsp is interpreted as time a sample path of the stochastic process X t w t T displaystyle X t omega t in T nbsp 50 This means that for a fixed w W displaystyle omega in Omega nbsp there exists a sample function that maps the index set T displaystyle T nbsp to the state space S displaystyle S nbsp 28 Other names for a sample function of a stochastic process include trajectory path function 140 or path 141 Increment edit An increment of a stochastic process is the difference between two random variables of the same stochastic process For a stochastic process with an index set that can be interpreted as time an increment is how much the stochastic process changes over a certain time period For example if X t t T displaystyle X t t in T nbsp is a stochastic process with state space S displaystyle S nbsp and index set T 0 displaystyle T 0 infty nbsp then for any two non negative numbers t 1 0 displaystyle t 1 in 0 infty nbsp and t 2 0 displaystyle t 2 in 0 infty nbsp such that t 1 t 2 displaystyle t 1 leq t 2 nbsp the difference X t 2 X t 1 displaystyle X t 2 X t 1 nbsp is a S displaystyle S nbsp valued random variable known as an increment 48 49 When interested in the increments often the state space S displaystyle S nbsp is the real line or the natural numbers but it can be n displaystyle n nbsp dimensional Euclidean space or more abstract spaces such as Banach spaces 49 Further definitions edit Law edit For a stochastic process X W S T displaystyle X colon Omega rightarrow S T nbsp defined on the probability space W F P displaystyle Omega mathcal F P nbsp the law of stochastic process X displaystyle X nbsp is defined as the image measure m P X 1 displaystyle mu P circ X 1 nbsp where P displaystyle P nbsp is a probability measure the symbol displaystyle circ nbsp denotes function composition and X 1 displaystyle X 1 nbsp is the pre image of the measurable function or equivalently the S T displaystyle S T nbsp valued random variable X displaystyle X nbsp where S T displaystyle S T nbsp is the space of all the possible S displaystyle S nbsp valued functions of t T displaystyle t in T nbsp so the law of a stochastic process is a probability measure 27 68 142 143 For a measurable subset B displaystyle B nbsp of S T displaystyle S T nbsp the pre image of X displaystyle X nbsp gives X 1 B w W X w B displaystyle X 1 B omega in Omega X omega in B nbsp so the law of a X displaystyle X nbsp can be written as 28 m B P w W X w B displaystyle mu B P omega in Omega X omega in B nbsp The law of a stochastic process or a random variable is also called the probability law probability distribution or the distribution 133 142 144 145 146 Finite dimensional probability distributions edit Main article Finite dimensional distribution For a stochastic process X displaystyle X nbsp with law m displaystyle mu nbsp its finite dimensional distribution for t 1 t n T displaystyle t 1 dots t n in T nbsp is defined as m t 1 t n P X t 1 X t n 1 displaystyle mu t 1 dots t n P circ X t 1 dots X t n 1 nbsp This measure m t 1 t n displaystyle mu t 1 t n nbsp is the joint distribution of the random vector X t 1 X t n displaystyle X t 1 dots X t n nbsp it can be viewed as a projection of the law m displaystyle mu nbsp onto a finite subset of T displaystyle T nbsp 27 147 For any measurable subset C displaystyle C nbsp of the n displaystyle n nbsp fold Cartesian power S n S S displaystyle S n S times dots times S nbsp the finite dimensional distributions of a stochastic process X displaystyle X nbsp can be written as 28 m t 1 t n C P w W X t 1 w X t n w C displaystyle mu t 1 dots t n C P Big big omega in Omega big X t 1 omega dots X t n omega big in C big Big nbsp The finite dimensional distributions of a stochastic process satisfy two mathematical conditions known as consistency conditions 57 Stationarity edit Main article Stationary process Stationarity is a mathematical property that a stochastic process has when all the random variables of that stochastic process are identically distributed In other words if X displaystyle X nbsp is a stationary stochastic process then for any t T displaystyle t in T nbsp the random variable X t displaystyle X t nbsp has the same distribution which means that for any set of n displaystyle n nbsp index set values t 1 t n displaystyle t 1 dots t n nbsp the corresponding n displaystyle n nbsp random variables X t 1 X t n displaystyle X t 1 dots X t n nbsp all have the same probability distribution The index set of a stationary stochastic process is usually interpreted as time so it can be the integers or the real line 148 149 But the concept of stationarity also exists for point processes and random fields where the index set is not interpreted as time 148 150 151 When the index set T displaystyle T nbsp can be interpreted as time a stochastic process is said to be stationary if its finite dimensional distributions are invariant under translations of time This type of stochastic process can be used to describe a physical system that is in steady state but still experiences random fluctuations 148 The intuition behind stationarity is that as time passes the distribution of the stationary stochastic process remains the same 152 A sequence of random variables forms a stationary stochastic process only if the random variables are identically distributed 148 A stochastic process with the above definition of stationarity is sometimes said to be strictly stationary but there are other forms of stationarity One example is when a discrete time or continuous time stochastic process X displaystyle X nbsp is said to be stationary in the wide sense then the process X displaystyle X nbsp has a finite second moment for all t T displaystyle t in T nbsp and the covariance of the two random variables X t displaystyle X t nbsp and X t h displaystyle X t h nbsp depends only on the number h displaystyle h nbsp for all t T displaystyle t in T nbsp 152 153 Khinchin introduced the related concept of stationarity in the wide sense which has other names including covariance stationarity or stationarity in the broad sense 153 154 Filtration edit A filtration is an increasing sequence of sigma algebras defined in relation to some probability space and an index set that has some total order relation such as in the case of the index set being some subset of the real numbers More formally if a stochastic process has an index set with a total order then a filtration F t t T displaystyle mathcal F t t in T nbsp on a probability space W F P displaystyle Omega mathcal F P nbsp is a family of sigma algebras such that F s F t F displaystyle mathcal F s subseteq mathcal F t subseteq mathcal F nbsp for all s t displaystyle s leq t nbsp where t s T displaystyle t s in T nbsp and displaystyle leq nbsp denotes the total order of the index set T displaystyle T nbsp 51 With the concept of a filtration it is possible to study the amount of information contained in a stochastic process X t displaystyle X t nbsp at t T displaystyle t in T nbsp which can be interpreted as time t displaystyle t nbsp 51 155 The intuition behind a filtration F t displaystyle mathcal F t nbsp is that as time t displaystyle t nbsp passes more and more information on X t displaystyle X t nbsp is known or available which is captured in F t displaystyle mathcal F t nbsp resulting in finer and finer partitions of W displaystyle Omega nbsp 156 157 Modification edit A modification of a stochastic process is another stochastic process which is closely related to the original stochastic process More precisely a stochastic process X displaystyle X nbsp that has the same index set T displaystyle T nbsp state space S displaystyle S nbsp and probability space W F P displaystyle Omega cal F P nbsp as another stochastic process Y displaystyle Y nbsp is said to be a modification of Y displaystyle Y nbsp if for all t T displaystyle t in T nbsp the following P X t Y t 1 displaystyle P X t Y t 1 nbsp holds Two stochastic processes that are modifications of each other have the same finite dimensional law 158 and they are said to be stochastically equivalent or equivalent 159 Instead of modification the term version is also used 150 160 161 162 however some authors use the term version when two stochastic processes have the same finite dimensional distributions but they may be defined on different probability spaces so two processes that are modifications of each other are also versions of each other in the latter sense but not the converse 163 142 If a continuous time real valued stochastic process meets certain moment conditions on its increments then the Kolmogorov continuity theorem says that there exists a modification of this process that has continuous sample paths with probability one so the stochastic process has a continuous modification or version 161 162 164 The theorem can also be generalized to random fields so the index set is n displaystyle n nbsp dimensional Euclidean space 165 as well as to stochastic processes with metric spaces as their state spaces 166 Indistinguishable edit Two stochastic processes X displaystyle X nbsp and Y displaystyle Y nbsp defined on the same probability space W F P displaystyle Omega mathcal F P nbsp with the same index set T displaystyle T nbsp and set space S displaystyle S nbsp are said be indistinguishable if the following P X t Y t for all t T 1 displaystyle P X t Y t text for all t in T 1 nbsp holds 142 158 If two X displaystyle X nbsp and Y displaystyle Y nbsp are modifications of each other and are almost surely continuous then X displaystyle X nbsp and Y displaystyle Y nbsp are indistinguishable 167 Separability edit Separability is a property of a stochastic process based on its index set in relation to the probability measure The property is assumed so that functionals of stochastic processes or random fields with uncountable index sets can form random variables For a stochastic process to be separable in addition to other conditions its index set must be a separable space b which means that the index set has a dense countable subset 150 168 More precisely a real valued continuous time stochastic process X displaystyle X nbsp with a probability space W F P displaystyle Omega cal F P nbsp is separable if its index set T displaystyle T nbsp has a dense countable subset U T displaystyle U subset T nbsp and there is a set W 0 W displaystyle Omega 0 subset Omega nbsp of probability zero so P W 0 0 displaystyle P Omega 0 0 nbsp such that for every open set G T displaystyle G subset T nbsp and every closed set F R displaystyle F subset textstyle R infty infty nbsp the two events X t F for all t G U displaystyle X t in F text for all t in G cap U nbsp and X t F for all t G displaystyle X t in F text for all t in G nbsp differ from each other at most on a subset of W 0 displaystyle Omega 0 nbsp 169 170 171 The definition of separability c can also be stated for other index sets and state spaces 174 such as in the case of random fields where the index set as well as the state space can be n displaystyle n nbsp dimensional Euclidean space 30 150 The concept of separability of a stochastic process was introduced by Joseph Doob 168 The underlying idea of separability is to make a countable set of points of the index set determine the properties of the stochastic process 172 Any stochastic process with a countable index set already meets the separability conditions so discrete time stochastic processes are always separable 175 A theorem by Doob sometimes known as Doob s separability theorem says that any real valued continuous time stochastic process has a separable modification 168 170 176 Versions of this theorem also exist for more general stochastic processes with index sets and state spaces other than the real line 136 Independence edit Two stochastic processes X displaystyle X nbsp and Y displaystyle Y nbsp defined on the same probability space W F P displaystyle Omega mathcal F P nbsp with the same index set T displaystyle T nbsp are said be independent if for all n N displaystyle n in mathbb N nbsp and for every choice of epochs t 1 t n T displaystyle t 1 ldots t n in T nbsp the random vectors X t 1 X t n displaystyle left X t 1 ldots X t n right nbsp and Y t 1 Y t n displaystyle left Y t 1 ldots Y t n right nbsp are independent 177 p 515 Uncorrelatedness edit Two stochastic processes X t displaystyle left X t right nbsp and Y t displaystyle left Y t right nbsp are called uncorrelated if their cross covariance K X Y t 1 t 2 E X t 1 m X t 1 Y t 2 m Y t 2 displaystyle operatorname K mathbf X mathbf Y t 1 t 2 operatorname E left left X t 1 mu X t 1 right left Y t 2 mu Y t 2 right right nbsp is zero for all times 178 p 142 Formally X t Y t uncorrelated K X Y t 1 t 2 0 t 1 t 2 displaystyle left X t right left Y t right text uncorrelated quad iff quad operatorname K mathbf X mathbf Y t 1 t 2 0 quad forall t 1 t 2 nbsp Independence implies uncorrelatedness edit If two stochastic processes X displaystyle X nbsp and Y displaystyle Y nbsp are independent then they are also uncorrelated 178 p 151 Orthogonality edit Two stochastic processes X t displaystyle left X t right nbsp and Y t displaystyle left Y t right nbsp are called orthogonal if their cross correlation R X Y t 1 t 2 E X t 1 Y t 2 displaystyle operatorname R mathbf X mathbf Y t 1 t 2 operatorname E X t 1 overline Y t 2 nbsp is zero for all times 178 p 142 Formally X t Y t orthogonal R X Y t 1 t 2 0 t 1 t 2 displaystyle left X t right left Y t right text orthogonal quad iff quad operatorname R mathbf X mathbf Y t 1 t 2 0 quad forall t 1 t 2 nbsp Skorokhod space edit Main article Skorokhod space A Skorokhod space also written as Skorohod space is a mathematical space of all the functions that are right continuous with left limits defined on some interval of the real line such as 0 1 displaystyle 0 1 nbsp or 0 displaystyle 0 infty nbsp and take values on the real line or on some metric space 179 180 181 Such functions are known as cadlag or cadlag functions based on the acronym of the French phrase continue a droite limite a gauche 179 182 A Skorokhod function space introduced by Anatoliy Skorokhod 181 is often denoted with the letter D displaystyle D nbsp 179 180 181 182 so the function space is also referred to as space D displaystyle D nbsp 179 183 184 The notation of this function space can also include the interval on which all the cadlag functions are defined so for example D 0 1 displaystyle D 0 1 nbsp denotes the space of cadlag functions defined on the unit interval 0 1 displaystyle 0 1 nbsp 182 184 185 Skorokhod function spaces are frequently used in the theory of stochastic processes because it often assumed that the sample functions of continuous time stochastic processes belong to a Skorokhod space 181 183 Such spaces contain continuous functions which correspond to sample functions of the Wiener process But the space also has functions with discontinuities which means that the sample functions of stochastic processes with jumps such as the Poisson process on the real line are also members of this space 184 186 Regularity edit In the context of mathematical construction of stochastic processes the term regularity is used when discussing and assuming certain conditions for a stochastic process to resolve possible construction issues 187 188 For example to study stochastic processes with uncountable index sets it is assumed that the stochastic process adheres to some type of regularity condition such as the sample functions being continuous 189 190 Further examples editMarkov processes and chains edit Main article Markov chain Markov processes are stochastic processes traditionally in discrete or continuous time that have the Markov property which means the next value of the Markov process depends on the current value but it is conditionally independent of the previous values of the stochastic process In other words the behavior of the process in the future is stochastically independent of its behavior in the past given the current state of the process 191 192 The Brownian motion process and the Poisson process in one dimension are both examples of Markov processes 193 in continuous time while random walks on the integers and the gambler s ruin problem are examples of Markov processes in discrete time 194 195 A Markov chain is a type of Markov process that has either discrete state space or discrete index set often representing time but the precise definition of a Markov chain varies 196 For example it is common to define a Markov chain as a Markov process in either discrete or continuous time with a countable state space thus regardless of the nature of time 197 198 199 200 but it has been also common to define a Markov chain as having discrete time in either countable or continuous state space thus regardless of the state space 196 It has been argued that the first definition of a Markov chain where it has discrete time now tends to be used despite the second definition having been used by researchers like Joseph Doob and Kai Lai Chung 201 Markov processes form an important class of stochastic processes and have applications in many areas 39 202 For example they are the basis for a general stochastic simulation method known as Markov chain Monte Carlo which is used for simulating random objects with specific probability distributions and has found application in Bayesian statistics 203 204 The concept of the Markov property was originally for stochastic processes in continuous and discrete time but the property has been adapted for other index sets such as n displaystyle n nbsp dimensional Euclidean space which results in collections of random variables known as Markov random fields 205 206 207 Martingale edit Main article Martingale probability theory A martingale is a discrete time or continuous time stochastic process with the property that at every instant given the current value and all the past values of the process the conditional expectation of every future value is equal to the current value In discrete time if this property holds for the next value then it holds for all future values The exact mathematical definition of a martingale requires two other conditions coupled with the mathematical concept of a filtration which is related to the intuition of increasing available information as time passes Martingales are usually defined to be real valued 208 209 155 but they can also be complex valued 210 or even more general 211 A symmetric random walk and a Wiener process with zero drift are both examples of martingales respectively in discrete and continuous time 208 209 For a sequence of independent and identically distributed random variables X 1 X 2 X 3 displaystyle X 1 X 2 X 3 dots nbsp with zero mean the stochastic process formed from the successive partial sums X 1 X 1 X 2 X 1 X 2 X 3 displaystyle X 1 X 1 X 2 X 1 X 2 X 3 dots nbsp is a discrete time martingale 212 In this aspect discrete time martingales generalize the idea of partial sums of independent random variables 213 Martingales can also be created from stochastic processes by applying some suitable transformations which is the case for the homogeneous Poisson process on the real line resulting in a martingale called the compensated Poisson process 209 Martingales can also be built from other martingales 212 For example there are martingales based on the martingale the Wiener process forming continuous time martingales 208 214 Martingales mathematically formalize the idea of a fair game where it is possible form reasonable expectations for payoffs 215 and they were originally developed to show that it is not possible to gain an unfair advantage in such a game 216 But now they are used in many areas of probability which is one of the main reasons for studying them 155 216 217 Many problems in probability have been solved by finding a martingale in the problem and studying it 218 Martingales will converge given some conditions on their moments so they are often used to derive convergence results due largely to martingale convergence theorems 213 219 220 Martingales have many applications in statistics but it has been remarked that its use and application are not as widespread as it could be in the field of statistics particularly statistical inference 221 They have found applications in areas in probability theory such as queueing theory and Palm calculus 222 and other fields such as economics 223 and finance 17 Levy process edit Main article Levy process Levy processes are types of stochastic processes that can be considered as generalizations of random walks in continuous time 49 224 These processes have many applications in fields such as finance fluid mechanics physics and biology 225 226 The main defining characteristics of these processes are their stationarity and independence properties so they were known as processes with stationary and independent increments In other words a stochastic process X displaystyle X nbsp is a Levy process if for n displaystyle n nbsp non negatives numbers 0 t 1 t n displaystyle 0 leq t 1 leq dots leq t n nbsp the corresponding n 1 displaystyle n 1 nbsp increments X t 2 X t 1 X t n X t n 1 displaystyle X t 2 X t 1 dots X t n X t n 1 nbsp are all independent of each other and the distribution of each increment only depends on the difference in time 49 A Levy process can be defined such that its state space is some abstract mathematical space such as a Banach space but the processes are often defined so that they take values in Euclidean space The index set is the non negative numbers so I 0 displaystyle I 0 infty nbsp which gives the interpretation of time Important stochastic processes such as the Wiener process the homogeneous Poisson process in one dimension and subordinators are all Levy processes 49 224 Random field edit Main article Random field A random field is a collection of random variables indexed by a n displaystyle n nbsp dimensional Euclidean space or some manifold In general a random field can be considered an example of a stochastic or random process where the index set is not necessarily a subset of the real line 30 But there is a convention that an indexed collection of random variables is called a random field when the index has two or more dimensions 5 28 227 If the specific definition of a stochastic process requires the index set to be a subset of the real line then the random field can be considered as a generalization of stochastic process 228 Point process edit Main article Point process A point process is a collection of points randomly located on some mathematical space such as the real line n displaystyle n nbsp dimensional Euclidean space or more abstract spaces Sometimes the term point process is not preferred as historically the word process denoted an evolution of some system in time so a point process is also called a random point field 229 There are different interpretations of a point process such a random counting measure or a random set 230 231 Some authors regard a point process and stochastic process as two different objects such that a point process is a random object that arises from or is associated with a stochastic process 232 233 though it has been remarked that the difference between point processes and stochastic processes is not clear 233 Other authors consider a point process as a stochastic process where the process is indexed by sets of the underlying space d on which it is defined such as the real line or n displaystyle n nbsp dimensional Euclidean space 236 237 Other stochastic processes such as renewal and counting processes are studied in the theory of point processes 238 233 History editEarly probability theory edit Probability theory has its origins in games of chance which have a long history with some games being played thousands of years ago 239 240 but very little analysis on them was done in terms of probability 239 241 The year 1654 is often considered the birth of probability theory when French mathematicians Pierre Fermat and Blaise Pascal had a written correspondence on probability motivated by a gambling problem 239 242 243 But there was earlier mathematical work done on the probability of gambling games such as Liber de Ludo Aleae by Gerolamo Cardano written in the 16th century but posthumously published later in 1663 239 244 After Cardano Jakob Bernoulli e wrote Ars Conjectandi which is considered a significant event in the history of probability theory 239 Bernoulli s book was published also posthumously in 1713 and inspired many mathematicians to study probability 239 246 247 But despite some renowned mathematicians contributing to probability theory such as Pierre Simon Laplace Abraham de Moivre Carl Gauss Simeon Poisson and Pafnuty Chebyshev 248 249 most of the mathematical community f did not consider probability theory to be part of mathematics until the 20th century 248 250 251 252 Statistical mechanics edit In the physical sciences scientists developed in the 19th century the discipline of statistical mechanics where physical systems such as containers filled with gases are regarded or treated mathematically as collections of many moving particles Although there were attempts to incorporate randomness into statistical physics by some scientists such as Rudolf Clausius most of the work had little or no randomness 253 254 This changed in 1859 when James Clerk Maxwell contributed significantly to the field more specifically to the kinetic theory of gases by presenting work where he modelled the gas particles as moving in random directions at random velocities 255 256 The kinetic theory of gases and statistical physics continued to be developed in the second half of the 19th century with work done chiefly by Clausius Ludwig Boltzmann and Josiah Gibbs which would later have an influence on Albert Einstein s mathematical model for Brownian movement 257 Measure theory and probability theory edit At the International Congress of Mathematicians in Paris in 1900 David Hilbert presented a list of mathematical problems where his sixth problem asked for a mathematical treatment of physics and probability involving axioms 249 Around the start of the 20th century mathematicians developed measure theory a branch of mathematics for studying integrals of mathematical functions where two of the founders were French mathematicians Henri Lebesgue and Emile Borel In 1925 another French mathematician Paul Levy published the first probability book that used ideas from measure theory 249 In the 1920s fundamental contributions to probability theory were made in the Soviet Union by mathematicians such as Sergei Bernstein Aleksandr Khinchin g and Andrei Kolmogorov 252 Kolmogorov published in 1929 his first attempt at presenting a mathematical foundation based on measure theory for probability theory 258 In the early 1930s Khinchin and Kolmogorov set up probability seminars which were attended by researchers such as Eugene Slutsky and Nikolai Smirnov 259 and Khinchin gave the first mathematical definition of a stochastic process as a set of random variables indexed by the real line 63 260 h Birth of modern probability theory edit In 1933 Andrei Kolmogorov published in German his book on the foundations of probability theory titled Grundbegriffe der Wahrscheinlichkeitsrechnung i where Kolmogorov used measure theory to develop an axiomatic framework for probability theory The publication of this book is now widely considered to be the birth of modern probability theory when the theories of probability and stochastic processes became parts of mathematics 249 252 After the publication of Kolmogorov s book further fundamental work on probability theory and stochastic processes was done by Khinchin and Kolmogorov as well as other mathematicians such as Joseph Doob William Feller Maurice Frechet Paul Levy Wolfgang Doeblin and Harald Cramer 249 252 Decades later Cramer referred to the 1930s as the heroic period of mathematical probability theory 252 World War II greatly interrupted the development of probability theory causing for example the migration of Feller from Sweden to the United States of America 252 and the death of Doeblin considered now a pioneer in stochastic processes 262 nbsp Mathematician Joseph Doob did early work on the theory of stochastic processes making fundamental contributions particularly in the theory of martingales 263 261 His book Stochastic Processes is considered highly influential in the field of probability theory 264 Stochastic processes after World War II edit After World War II the study of probability theory and stochastic processes gained more attention from mathematicians with significant contributions made in many areas of probability and mathematics as well as the creation of new areas 252 265 Starting in the 1940s Kiyosi Ito published papers developing the field of stochastic calculus which involves stochastic integrals and stochastic differential equations based on the Wiener or Brownian motion process 266 Also starting in the 1940s connections were made between stochastic processes particularly martingales and the mathematical field of potential theory with early ideas by Shizuo Kakutani and then later work by Joseph Doob 265 Further work considered pioneering was done by Gilbert Hunt in the 1950s connecting Markov processes and potential theory which had a significant effect on the theory of Levy processes and led to more interest in studying Markov processes with methods developed by Ito 21 267 268 In 1953 Doob published his book Stochastic processes which had a strong influence on the theory of stochastic processes and stressed the importance of measure theory in probability 265 264 Doob also chiefly developed the theory of martingales with later substantial contributions by Paul Andre Meyer Earlier work had been carried out by Sergei Bernstein Paul Levy and Jean Ville the latter adopting the term martingale for the stochastic process 269 270 Methods from the theory of martingales became popular for solving various probability problems Techniques and theory were developed to study Markov processes and then applied to martingales Conversely methods from the theory of martingales were established to treat Markov processes 265 Other fields of probability were developed and used to study stochastic processes with one main approach being the theory of large deviations 265 The theory has many applications in statistical physics among other fields and has core ideas going back to at least the 1930s Later in the 1960s and 1970s fundamental work was done by Alexander Wentzell in the Soviet Union and Monroe D Donsker and Srinivasa Varadhan in the United States of America 271 which would later result in Varadhan winning the 2007 Abel Prize 272 In the 1990s and 2000s the theories of Schramm Loewner evolution 273 and rough paths 142 were introduced and developed to study stochastic processes and other mathematical objects in probability theory which respectively resulted in Fields Medals being awarded to Wendelin Werner 274 in 2008 and to Martin Hairer in 2014 275 The theory of stochastic processes still continues to be a focus of research with yearly international conferences on the topic of stochastic processes 45 225 Discoveries of specific stochastic processes edit Although Khinchin gave mathematical definitions of stochastic processes in the 1930s 63 260 specific stochastic processes had already been discovered in different settings such as the Brownian motion process and the Poisson process 21 24 Some families of stochastic processes such as point processes or renewal processes have long and complex histories stretching back centuries 276 Bernoulli process edit The Bernoulli process which can serve as a mathematical model for flipping a biased coin is possibly the first stochastic process to have been studied 81 The process is a sequence of independent Bernoulli trials 82 which are named after Jackob Bernoulli who used them to study games of chance including probability problems proposed and studied earlier by Christiaan Huygens 277 Bernoulli s work including the Bernoulli process were published in his book Ars Conjectandi in 1713 278 Random walks edit In 1905 Karl Pearson coined the term random walk while posing a problem describing a random walk on the plane which was motivated by an application in biology but such problems involving random walks had already been studied in other fields Certain gambling problems that were studied centuries earlier can be considered as problems involving random walks 89 278 For example the problem known as the Gambler s ruin is based on a simple random walk 195 279 and is an example of a random walk with absorbing barriers 242 280 Pascal Fermat and Huyens all gave numerical solutions to this problem without detailing their methods 281 and then more detailed solutions were presented by Jakob Bernoulli and Abraham de Moivre 282 For random walks in n displaystyle n nbsp dimensional integer lattices George Polya published in 1919 and 1921 work where he studied the probability of a symmetric random walk returning to a previous position in the lattice Polya showed that a symmetric random walk which has an equal probability to advance in any direction in the lattice will return to a previous position in the lattice an infinite number of times with probability one in one and two dimensions but with probability zero in three or higher dimensions 283 284 Wiener process edit The Wiener process or Brownian motion process has its origins in different fields including statistics finance and physics 21 In 1880 Danish astronomer Thorvald Thiele wrote a paper on the method of least squares where he used the process to study the errors of a model in time series analysis 285 286 287 The work is now considered as an early discovery of the statistical method known as Kalman filtering but the work was largely overlooked It is thought that the ideas in Thiele s paper were too advanced to have been understood by the broader mathematical and statistical community at the time 287 nbsp Norbert Wiener gave the first mathematical proof of the existence of the Wiener process This mathematical object had appeared previously in the work of Thorvald Thiele Louis Bachelier and Albert Einstein 21 The French mathematician Louis Bachelier used a Wiener process in his 1900 thesis 288 289 in order to model price changes on the Paris Bourse a stock exchange 290 without knowing the work of Thiele 21 It has been speculated that Bachelier drew ideas from the random walk model of Jules Regnault but Bachelier did not cite him 291 and Bachelier s thesis is now considered pioneering in the field of financial mathematics 290 291 It is commonly thought that Bachelier s work gained little attention and was forgotten for decades until it was rediscovered in the 1950s by the Leonard Savage and then become more popular after Bachelier s thesis was translated into English in 1964 But the work was never forgotten in the mathematical community as Bachelier published a book in 1912 detailing his ideas 291 which was cited by mathematicians including Doob Feller 291 and Kolmogorov 21 The book continued to be cited but then starting in the 1960s the original thesis by Bachelier began to be cited more than his book when economists started citing Bachelier s work 291 In 1905 Albert Einstein published a paper where he studied the physical observation of Brownian motion or movement to explain the seemingly random movements of particles in liquids by using ideas from the kinetic theory of gases Einstein derived a differential equation known as a diffusion equation for describing the probability of finding a particle in a certain region of space Shortly after Einstein s first paper on Brownian movement Marian Smoluchowski published work where he cited Einstein but wrote that he had independently derived the equivalent results by using a different method 292 Einstein s work as well as experimental results obtained by Jean Perrin later inspired Norbert Wiener in the 1920s 293 to use a type of measure theory developed by Percy Daniell and Fourier analysis to prove the existence of the Wiener process as a mathematical object 21 Poisson process edit The Poisson process is named after Simeon Poisson due to its definition involving the Poisson distribution but Poisson never studied the process 22 294 There are a number of claims for early uses or discoveries of the Poisson process 22 24 At the beginning of the 20th century the Poisson process would arise independently in different situations 22 24 In Sweden 1903 Filip Lundberg published a thesis containing work now considered fundamental and pioneering where he proposed to model insurance claims with a homogeneous Poisson process 295 296 Another discovery occurred in Denmark in 1909 when A K Erlang derived the Poisson distribution when developing a mathematical model for the number of incoming phone calls in a finite time interval Erlang was not at the time aware of Poisson s earlier work and assumed that the number phone calls arriving in each interval of time were independent to each other He then found the limiting case which is effectively recasting the Poisson distribution as a limit of the binomial distribution 22 In 1910 Ernest Rutherford and Hans Geiger published experimental results on counting alpha particles Motivated by their work Harry Bateman studied the counting problem and derived Poisson probabilities as a solution to a family of differential equations resulting in the independent discovery of the Poisson process 22 After this time there were many studies and applications of the Poisson process but its early history is complicated which has been explained by the various applications of the process in numerous fields by biologists ecologists engineers and various physical scientists 22 Markov processes edit Markov processes and Markov chains are named after Andrey Markov who studied Markov chains in the early 20th century Markov was interested in studying an extension of independent random sequences In his first paper on Markov chains published in 1906 Markov showed that under certain conditions the average outcomes of the Markov chain would converge to a fixed vector of values so proving a weak law of large numbers without the independence assumption 297 298 299 which had been commonly regarded as a requirement for such mathematical laws to hold 299 Markov later used Markov chains to study the distribution of vowels in Eugene Onegin written by Alexander Pushkin and proved a central limit theorem for such chains 300 297 In 1912 Poincare studied Markov chains on finite groups with an aim to study card shuffling Other early uses of Markov chains include a diffusion model introduced by Paul and Tatyana Ehrenfest in 1907 and a branching process introduced by Francis Galton and Henry William Watson in 1873 preceding the work of Markov 297 298 After the work of Galton and Watson it was later revealed that their branching process had been independently discovered and studied around three decades earlier by Irenee Jules Bienayme 301 Starting in 1928 Maurice Frechet became interested in Markov chains eventually resulting in him publishing in 1938 a detailed study on Markov chains 297 302 Andrei Kolmogorov developed in a 1931 paper a large part of the early theory of continuous time Markov processes 252 258 Kolmogorov was partly inspired by Louis Bachelier s 1900 work on fluctuations in the stock market as well as Norbert Wiener s work on Einstein s model of Brownian movement 258 303 He introduced and studied a particular set of Markov processes known as diffusion processes where he derived a set of differential equations describing the processes 258 304 Independent of Kolmogorov s work Sydney Chapman derived in a 1928 paper an equation now called the Chapman Kolmogorov equation in a less mathematically rigorous way than Kolmogorov while studying Brownian movement 305 The differential equations are now called the Kolmogorov equations 306 or the Kolmogorov Chapman equations 307 Other mathematicians who contributed significantly to the foundations of Markov processes include William Feller starting in the 1930s and then later Eugene Dynkin starting in the 1950s 252 Levy processes edit Levy processes such as the Wiener process and the Poisson process on the real line are named after Paul Levy who started studying them in the 1930s 225 but they have connections to infinitely divisible distributions going back to the 1920s 224 In a 1932 paper Kolmogorov derived a characteristic function for random variables associated with Levy processes This result was later derived under more general conditions by Levy in 1934 and then Khinchin independently gave an alternative form for this characteristic function in 1937 252 308 In addition to Levy Khinchin and Kolomogrov early fundamental contributions to the theory of Levy processes were made by Bruno de Finetti and Kiyosi Ito 224 Mathematical construction editIn mathematics constructions of mathematical objects are needed which is also the case for stochastic processes to prove that they exist mathematically 57 There are two main approaches for constructing a stochastic process One approach involves considering a measurable space of functions defining a suitable measurable mapping from a probability space to this measurable space of functions and then deriving the corresponding finite dimensional distributions 309 Another approach involves defining a collection of random variables to have specific finite dimensional distributions and then using Kolmogorov s existence theorem j to prove a corresponding stochastic process exists 57 309 This theorem which is an existence theorem for measures on infinite product spaces 313 says that if any finite dimensional distributions satisfy two conditions known as consistency conditions then there exists a stochastic process with those finite dimensional distributions 57 Construction issues edit When constructing continuous time stochastic processes certain mathematical difficulties arise due to the uncountable index sets which do not occur with discrete time processes 58 59 One problem is that is it possible to have more than one stochastic process with the same finite dimensional distributions For example both the left continuous modification and the right continuous modification of a Poisson process have the same finite dimensional distributions 314 This means that the distribution of the stochastic process does not necessarily specify uniquely the properties of the sample functions of the stochastic process 309 315 Another problem is that functionals of continuous time process that rely upon an uncountable number of points of the index set may not be measurable so the probabilities of certain events may not be well defined 168 For example the supremum of a stochastic process or random field is not necessarily a well defined random variable 30 59 For a continuous time stochastic process X displaystyle X nbsp other characteristics that depend on an uncountable number of points of the index set T displaystyle T nbsp include 168 a sample function of a stochastic process X displaystyle X nbsp is a continuous function of t T displaystyle t in T nbsp a sample function of a stochastic process X displaystyle X nbsp is a bounded function of t T displaystyle t in T nbsp and a sample function of a stochastic process X displaystyle X nbsp is an increasing function of t T displaystyle t in T nbsp To overcome these two difficulties different assumptions and approaches are possible 69 Resolving construction issues edit One approach for avoiding mathematical construction issues of stochastic processes proposed by Joseph Doob is to assume that the stochastic process is separable 316 Separability ensures that infinite dimensional distributions determine the properties of sample functions by requiring that sample functions are essentially determined by their values on a dense countable set of points in the index set 317 Furthermore if a stochastic process is separable then functionals of an uncountable number of points of the index set are measurable and their probabilities can be studied 168 317 Another approach is possible originally developed by Anatoliy Skorokhod and Andrei Kolmogorov 318 for a continuous time stochastic process with any metric space as its state space For the construction of such a stochastic process it is assumed that the sample functions of the stochastic process belong to some suitable function space which is usually the Skorokhod space consisting of all right continuous functions with left limits This approach is now more used than the separability assumption 69 263 but such a stochastic process based on this approach will be automatically separable 319 Although less used the separability assumption is considered more general because every stochastic process has a separable version 263 It is also used when it is not possible to construct a stochastic process in a Skorokhod space 173 For example separability is assumed when constructing and studying random fields where the collection of random variables is now indexed by sets other than the real line such as n displaystyle n nbsp dimensional Euclidean space 30 320 See also editList of stochastic processes topics Covariance function Deterministic system Dynamics of Markovian particles Entropy rate for a stochastic process Ergodic process Gillespie algorithm Interacting particle system Law stochastic processes Markov chain Stochastic cellular automaton Random field Randomness Stationary process Statistical model Stochastic calculus Stochastic control Stochastic parrot Stochastic processes and boundary value problemsNotes edit The term Brownian motion can refer to the physical process also known as Brownian movement and the stochastic process a mathematical object but to avoid ambiguity this article uses the terms Brownian motion process or Wiener process for the latter in a style similar to for example Gikhman and Skorokhod 19 or Rosenblatt 20 The term separable appears twice here with two different meanings where the first meaning is from probability and the second from topology and analysis For a stochastic process to be separable in a probabilistic sense its index set must be a separable space in a topological or analytic sense in addition to other conditions 136 The definition of separability for a continuous time real valued stochastic process can be stated in other ways 172 173 In the context of point processes the term state space can mean the space on which the point process is defined such as the real line 234 235 which corresponds to the index set in stochastic process terminology Also known as James or Jacques Bernoulli 245 It has been remarked that a notable exception was the St Petersburg School in Russia where mathematicians led by Chebyshev studied probability theory 250 The name Khinchin is also written in or transliterated into English as Khintchine 63 Doob when citing Khinchin uses the term chance variable which used to be an alternative term for random variable 261 Later translated into English and published in 1950 as Foundations of the Theory of Probability 249 The theorem has other names including Kolmogorov s consistency theorem 310 Kolmogorov s extension theorem 311 or the Daniell Kolmogorov theorem 312 References edit a b c d e f g h i Joseph L Doob 1990 Stochastic processes Wiley pp 46 47 a b c d L C G Rogers David Williams 2000 Diffusions Markov Processes and Martingales Volume 1 Foundations Cambridge University Press p 1 ISBN 978 1 107 71749 7 a b c J Michael Steele 2012 Stochastic Calculus and Financial Applications Springer Science amp Business Media p 29 ISBN 978 1 4684 9305 4 a b c d e Emanuel Parzen 2015 Stochastic Processes Courier Dover Publications pp 7 8 ISBN 978 0 486 79688 8 a b c d e f g h i j k l Iosif Ilyich Gikhman Anatoly Vladimirovich Skorokhod 1969 Introduction to the Theory of Random Processes Courier Corporation p 1 ISBN 978 0 486 69387 3 Bressloff Paul C 2014 Stochastic Processes in Cell Biology Springer ISBN 978 3 319 08488 6 Van Kampen N G 2011 Stochastic Processes in Physics and Chemistry Elsevier ISBN 978 0 08 047536 3 Lande Russell Engen Steinar Saether Bernt Erik 2003 Stochastic Population Dynamics in Ecology and Conservation Oxford University Press ISBN 978 0 19 852525 7 Laing Carlo Lord Gabriel J 2010 Stochastic Methods in Neuroscience Oxford University Press ISBN 978 0 19 923507 0 Paul Wolfgang Baschnagel Jorg 2013 Stochastic Processes From Physics to Finance Springer Science Business Media ISBN 978 3 319 00327 6 Dougherty Edward R 1999 Random processes for image and signal processing SPIE Optical Engineering Press ISBN 978 0 8194 2513 3 Bertsekas Dimitri P 1996 Stochastic Optimal Control The Discrete Time Case Athena Scientific ISBN 1 886529 03 5 Thomas M Cover Joy A Thomas 2012 Elements of Information Theory John Wiley amp Sons p 71 ISBN 978 1 118 58577 1 Baron Michael 2015 Probability and Statistics for Computer Scientists 2nd ed CRC Press p 131 ISBN 978 1 4987 6060 7 Baccelli Francois Blaszczyszyn Bartlomiej 2009 Stochastic Geometry and Wireless Networks Now Publishers Inc ISBN 978 1 60198 264 3 Steele J Michael 2001 Stochastic Calculus and Financial Applications Springer Science Business Media ISBN 978 0 387 95016 7 a b Musiela Marek Rutkowski Marek 2006 Martingale Methods in Financial Modelling Springer Science Business Media ISBN 978 3 540 26653 2 Shreve Steven E 2004 Stochastic Calculus for Finance II Continuous Time Models Springer Science Business Media ISBN 978 0 387 40101 0 Iosif Ilyich Gikhman Anatoly Vladimirovich Skorokhod 1969 Introduction to the Theory of Random Processes Courier Corporation ISBN 978 0 486 69387 3 Murray Rosenblatt 1962 Random Processes Oxford University Press a b c d e f g h i Jarrow Robert Protter Philip 2004 A short history of stochastic integration and mathematical finance the early years 1880 1970 A Festschrift for Herman Rubin Institute of Mathematical Statistics Lecture Notes Monograph Series pp 75 80 CiteSeerX 10 1 1 114 632 doi 10 1214 lnms 1196285381 ISBN 978 0 940600 61 4 ISSN 0749 2170 a b c d e f g h Stirzaker David 2000 Advice to Hedgehogs or Constants Can Vary The Mathematical Gazette 84 500 197 210 doi 10 2307 3621649 ISSN 0025 5572 JSTOR 3621649 S2CID 125163415 Donald L Snyder Michael I Miller 2012 Random Point Processes in Time and Space Springer Science amp Business Media p 32 ISBN 978 1 4612 3166 0 a b c d Guttorp Peter Thorarinsdottir Thordis L 2012 What Happened to Discrete Chaos the Quenouille Process and the Sharp Markov Property Some History of Stochastic Point Processes International Statistical Review 80 2 253 268 doi 10 1111 j 1751 5823 2012 00181 x ISSN 0306 7734 S2CID 80836 Gusak Dmytro Kukush Alexander Kulik Alexey Mishura Yuliya Pilipenko Andrey 2010 Theory of Stochastic Processes With Applications to Financial Mathematics and Risk Theory Springer Science amp Business Media p 21 ISBN 978 0 387 87862 1 Valeriy Skorokhod 2005 Basic Principles and Applications of Probability Theory Springer Science amp Business Media p 42 ISBN 978 3 540 26312 8 a b c d e f Olav Kallenberg 2002 Foundations of Modern Probability Springer Science amp Business Media pp 24 25 ISBN 978 0 387 95313 7 a b c d e f g h i j k l m n o p John Lamperti 1977 Stochastic processes a survey of the mathematical theory Springer Verlag pp 1 2 ISBN 978 3 540 90275 1 a b c d Loic Chaumont Marc Yor 2012 Exercises in Probability A Guided Tour from Measure Theory to Random Processes Via Conditioning Cambridge University Press p 175 ISBN 978 1 107 60655 5 a b c d e f g h Robert J Adler Jonathan E Taylor 2009 Random Fields and Geometry Springer Science amp Business Media pp 7 8 ISBN 978 0 387 48116 6 Gregory F Lawler Vlada Limic 2010 Random Walk A Modern Introduction Cambridge University Press ISBN 978 1 139 48876 1 David Williams 1991 Probability with Martingales Cambridge University Press ISBN 978 0 521 40605 5 L C G Rogers David Williams 2000 Diffusions Markov Processes and Martingales Volume 1 Foundations Cambridge University Press ISBN 978 1 107 71749 7 David Applebaum 2004 Levy Processes and Stochastic Calculus Cambridge University Press ISBN 978 0 521 83263 2 Mikhail Lifshits 2012 Lectures on Gaussian Processes Springer Science amp Business Media ISBN 978 3 642 24939 6 Robert J Adler 2010 The Geometry of Random Fields SIAM ISBN 978 0 89871 693 1 Samuel Karlin Howard E Taylor 2012 A First Course in Stochastic Processes Academic Press ISBN 978 0 08 057041 9 Bruce Hajek 2015 Random Processes for Engineers Cambridge University Press ISBN 978 1 316 24124 0 a b G Latouche V Ramaswami 1999 Introduction to Matrix Analytic Methods in Stochastic Modeling SIAM ISBN 978 0 89871 425 8 D J Daley David Vere Jones 2007 An Introduction to the Theory of Point Processes Volume II General Theory and Structure Springer Science amp Business Media ISBN 978 0 387 21337 8 Patrick Billingsley 2008 Probability and Measure Wiley India Pvt Limited ISBN 978 81 265 1771 8 Pierre Bremaud 2014 Fourier Analysis and Stochastic Processes Springer ISBN 978 3 319 09590 5 Adam Bobrowski 2005 Functional Analysis for Probability and Stochastic Processes An Introduction Cambridge University Press ISBN 978 0 521 83166 6 Applebaum David 2004 Levy processes From probability to finance and quantum groups Notices of the AMS 51 11 1336 1347 a b Jochen Blath Peter Imkeller Sylvie Roelly 2011 Surveys in Stochastic Processes European Mathematical Society ISBN 978 3 03719 072 2 Michel Talagrand 2014 Upper and Lower Bounds for Stochastic Processes Modern Methods and Classical Problems Springer Science amp Business Media pp 4 ISBN 978 3 642 54075 2 Paul C Bressloff 2014 Stochastic Processes in Cell Biology Springer pp vii ix ISBN 978 3 319 08488 6 a b c d Samuel Karlin Howard E Taylor 2012 A First Course in Stochastic Processes Academic Press p 27 ISBN 978 0 08 057041 9 a b c d e f g h i j Applebaum David 2004 Levy processes From probability to finance and quantum groups Notices of the AMS 51 11 1337 a b L C G Rogers David Williams 2000 Diffusions Markov Processes and Martingales Volume 1 Foundations Cambridge University Press pp 121 124 ISBN 978 1 107 71749 7 a b c d e f Ionut Florescu 2014 Probability and Stochastic Processes John Wiley amp Sons pp 294 295 ISBN 978 1 118 59320 2 a b Samuel Karlin Howard E Taylor 2012 A First Course in Stochastic Processes Academic Press p 26 ISBN 978 0 08 057041 9 Donald L Snyder Michael I Miller 2012 Random Point Processes in Time and Space Springer Science amp Business Media pp 24 25 ISBN 978 1 4612 3166 0 a b Patrick Billingsley 2008 Probability and Measure Wiley India Pvt Limited p 482 ISBN 978 81 265 1771 8 a b Alexander A Borovkov 2013 Probability Theory Springer Science amp Business Media p 527 ISBN 978 1 4471 5201 9 a b c Pierre Bremaud 2014 Fourier Analysis and Stochastic Processes Springer p 120 ISBN 978 3 319 09590 5 a b c d e Jeffrey S Rosenthal 2006 A First Look at Rigorous Probability Theory World Scientific Publishing Co Inc pp 177 178 ISBN 978 981 310 165 4 a b Peter E Kloeden Eckhard Platen 2013 Numerical Solution of Stochastic Differential Equations Springer Science amp Business Media p 63 ISBN 978 3 662 12616 5 a b c Davar Khoshnevisan 2006 Multiparameter Processes An Introduction to Random Fields Springer Science amp Business Media pp 153 155 ISBN 978 0 387 21631 7 a b Stochastic Oxford English Dictionary Online ed Oxford University Press Subscription or participating institution membership required O B Sheĭnin 2006 Theory of probability and statistics as exemplified in short dictums NG Verlag p 5 ISBN 978 3 938417 40 9 Oscar Sheynin Heinrich Strecker 2011 Alexandr A Chuprov Life Work Correspondence V amp R unipress GmbH p 136 ISBN 978 3 89971 812 6 a b c d Doob Joseph 1934 Stochastic Processes and Statistics Proceedings of the National Academy of Sciences of the United States of America 20 6 376 379 Bibcode 1934PNAS 20 376D doi 10 1073 pnas 20 6 376 PMC 1076423 PMID 16587907 Khintchine A 1934 Korrelationstheorie der stationeren stochastischen Prozesse Mathematische Annalen 109 1 604 615 doi 10 1007 BF01449156 ISSN 0025 5831 S2CID 122842868 span, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.