fbpx
Wikipedia

Random variable

A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events.[1] The term 'random variable' in its mathematical definition refers to neither randomness nor variability[2] but instead is a mathematical function in which

  • the domain is the set of possible outcomes in a sample space (e.g. the set which are the possible upper sides of a flipped coin heads or tails as the result from tossing a coin); and
  • the range is a measurable space (e.g. corresponding to the domain above, the range might be the set if say heads mapped to -1 and mapped to 1). Typically, the range of a random variable is set of real numbers.
This graph shows how random variable is a function from all possible outcomes to real values. It also shows how random variable is used for defining probability mass functions.

Informally, randomness typically represents some fundamental element of chance, such as in the roll of a die; it may also represent uncertainty, such as measurement error.[1] However, the interpretation of probability is philosophically complicated, and even in specific cases is not always straightforward. The purely mathematical analysis of random variables is independent of such interpretational difficulties, and can be based upon a rigorous axiomatic setup.

In the formal mathematical language of measure theory, a random variable is defined as a measurable function from a probability measure space (called the sample space) to a measurable space. This allows consideration of the pushforward measure, which is called the distribution of the random variable; the distribution is thus a probability measure on the set of all possible values of the random variable. It is possible for two random variables to have identical distributions but to differ in significant ways; for instance, they may be independent.

It is common to consider the special cases of discrete random variables and absolutely continuous random variables, corresponding to whether a random variable is valued in a countable subset or in an interval of real numbers. There are other important possibilities, especially in the theory of stochastic processes, wherein it is natural to consider random sequences or random functions. Sometimes a random variable is taken to be automatically valued in the real numbers, with more general random quantities instead being called random elements.

According to George Mackey, Pafnuty Chebyshev was the first person "to think systematically in terms of random variables".[3]

Definition edit

A random variable   is a measurable function   from a sample space   as a set of possible outcomes to a measurable space  . The technical axiomatic definition requires the sample space   to be a sample space of a probability triple   (see the measure-theoretic definition). A random variable is often denoted by capital Roman letters such as  .[4]

The probability that   takes on a value in a measurable set   is written as

 .

Standard case edit

In many cases,   is real-valued, i.e.  . In some contexts, the term random element (see extensions) is used to denote a random variable not of this form.

When the image (or range) of   is finitely or infinitely countable, the random variable is called a discrete random variable[5]: 399  and its distribution is a discrete probability distribution, i.e. can be described by a probability mass function that assigns a probability to each value in the image of  . If the image is uncountably infinite (usually an interval) then   is called a continuous random variable.[6][7] In the special case that it is absolutely continuous, its distribution can be described by a probability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous.[8]

Any random variable can be described by its cumulative distribution function, which describes the probability that the random variable will be less than or equal to a certain value.

Extensions edit

The term "random variable" in statistics is traditionally limited to the real-valued case ( ). In this case, the structure of the real numbers makes it possible to define quantities such as the expected value and variance of a random variable, its cumulative distribution function, and the moments of its distribution.

However, the definition above is valid for any measurable space   of values. Thus one can consider random elements of other sets  , such as random Boolean values, categorical values, complex numbers, vectors, matrices, sequences, trees, sets, shapes, manifolds, and functions. One may then specifically refer to a random variable of type  , or an  -valued random variable.

This more general concept of a random element is particularly useful in disciplines such as graph theory, machine learning, natural language processing, and other fields in discrete mathematics and computer science, where one is often interested in modeling the random variation of non-numerical data structures. In some cases, it is nonetheless convenient to represent each element of  , using one or more real numbers. In this case, a random element may optionally be represented as a vector of real-valued random variables (all defined on the same underlying probability space  , which allows the different random variables to covary). For example:

  • A random word may be represented as a random integer that serves as an index into the vocabulary of possible words. Alternatively, it can be represented as a random indicator vector, whose length equals the size of the vocabulary, where the only values of positive probability are  ,  ,   and the position of the 1 indicates the word.
  • A random sentence of given length   may be represented as a vector of   random words.
  • A random graph on   given vertices may be represented as a   matrix of random variables, whose values specify the adjacency matrix of the random graph.
  • A random function   may be represented as a collection of random variables  , giving the function's values at the various points   in the function's domain. The   are ordinary real-valued random variables provided that the function is real-valued. For example, a stochastic process is a random function of time, a random vector is a random function of some index set such as  , and random field is a random function on any set (typically time, space, or a discrete set).

Distribution functions edit

If a random variable   defined on the probability space   is given, we can ask questions like "How likely is it that the value of   is equal to 2?". This is the same as the probability of the event   which is often written as   or   for short.

Recording all these probabilities of outputs of a random variable   yields the probability distribution of  . The probability distribution "forgets" about the particular probability space used to define   and only records the probabilities of various output values of  . Such a probability distribution, if   is real-valued, can always be captured by its cumulative distribution function

 

and sometimes also using a probability density function,  . In measure-theoretic terms, we use the random variable   to "push-forward" the measure   on   to a measure   on  . The measure   is called the "(probability) distribution of  " or the "law of  ". [9] The density  , the Radon–Nikodym derivative of   with respect to some reference measure   on   (often, this reference measure is the Lebesgue measure in the case of continuous random variables, or the counting measure in the case of discrete random variables). The underlying probability space   is a technical device used to guarantee the existence of random variables, sometimes to construct them, and to define notions such as correlation and dependence or independence based on a joint distribution of two or more random variables on the same probability space. In practice, one often disposes of the space   altogether and just puts a measure on   that assigns measure 1 to the whole real line, i.e., one works with probability distributions instead of random variables. See the article on quantile functions for fuller development.

Examples edit

Discrete random variable edit

Consider an experiment where a person is chosen at random. An example of a random variable may be the person's height. Mathematically, the random variable is interpreted as a function which maps the person to his or her height. Associated with the random variable is a probability distribution that allows the computation of the probability that the height is in any subset of possible values, such as the probability that the height is between 180 and 190 cm, or the probability that the height is either less than 150 or more than 200 cm.

Another random variable may be the person's number of children; this is a discrete random variable with non-negative integer values. It allows the computation of probabilities for individual integer values – the probability mass function (PMF) – or for sets of values, including infinite sets. For example, the event of interest may be "an even number of children". For both finite and infinite event sets, their probabilities can be found by adding up the PMFs of the elements; that is, the probability of an even number of children is the infinite sum  .

In examples such as these, the sample space is often suppressed, since it is mathematically hard to describe, and the possible values of the random variables are then treated as a sample space. But when two random variables are measured on the same sample space of outcomes, such as the height and number of children being computed on the same random persons, it is easier to track their relationship if it is acknowledged that both height and number of children come from the same random person, for example so that questions of whether such random variables are correlated or not can be posed.

If   are countable sets of real numbers,   and  , then   is a discrete distribution function. Here   for  ,   for  . Taking for instance an enumeration of all rational numbers as   , one gets a discrete function that is not necessarily a step function (piecewise constant).

Coin toss edit

The possible outcomes for one coin toss can be described by the sample space  . We can introduce a real-valued random variable   that models a $1 payoff for a successful bet on heads as follows:

 

If the coin is a fair coin, Y has a probability mass function   given by:

 

Dice roll edit

 
If the sample space is the set of possible numbers rolled on two dice, and the random variable of interest is the sum S of the numbers on the two dice, then S is a discrete random variable whose distribution is described by the probability mass function plotted as the height of picture columns here.

A random variable can also be used to describe the process of rolling dice and the possible outcomes. The most obvious representation for the two-dice case is to take the set of pairs of numbers n1 and n2 from {1, 2, 3, 4, 5, 6} (representing the numbers on the two dice) as the sample space. The total number rolled (the sum of the numbers in each pair) is then a random variable X given by the function that maps the pair to the sum:

 
and (if the dice are fair) has a probability mass function fX given by:
 

Continuous random variable edit

Formally, a continuous random variable is a random variable whose cumulative distribution function is continuous everywhere.[10] There are no "gaps", which would correspond to numbers which have a finite probability of occurring. Instead, continuous random variables almost never take an exact prescribed value c (formally,  ) but there is a positive probability that its value will lie in particular intervals which can be arbitrarily small. Continuous random variables usually admit probability density functions (PDF), which characterize their CDF and probability measures; such distributions are also called absolutely continuous; but some continuous distributions are singular, or mixes of an absolutely continuous part and a singular part.

An example of a continuous random variable would be one based on a spinner that can choose a horizontal direction. Then the values taken by the random variable are directions. We could represent these directions by North, West, East, South, Southeast, etc. However, it is commonly more convenient to map the sample space to a random variable which takes values which are real numbers. This can be done, for example, by mapping a direction to a bearing in degrees clockwise from North. The random variable then takes values which are real numbers from the interval [0, 360), with all parts of the range being "equally likely". In this case, X = the angle spun. Any real number has probability zero of being selected, but a positive probability can be assigned to any range of values. For example, the probability of choosing a number in [0, 180] is 12. Instead of speaking of a probability mass function, we say that the probability density of X is 1/360. The probability of a subset of [0, 360) can be calculated by multiplying the measure of the set by 1/360. In general, the probability of a set for a given continuous random variable can be calculated by integrating the density over the given set.

More formally, given any interval  , a random variable   is called a "continuous uniform random variable" (CURV) if the probability that it takes a value in a subinterval depends only on the length of the subinterval. This implies that the probability of   falling in any subinterval   is proportional to the length of the subinterval, that is, if acdb, one has

 

where the last equality results from the unitarity axiom of probability. The probability density function of a CURV   is given by the indicator function of its interval of support normalized by the interval's length:

 
Of particular interest is the uniform distribution on the unit interval  . Samples of any desired probability distribution   can be generated by calculating the quantile function of   on a randomly-generated number distributed uniformly on the unit interval. This exploits properties of cumulative distribution functions, which are a unifying framework for all random variables.

Mixed type edit

A mixed random variable is a random variable whose cumulative distribution function is neither discrete nor everywhere-continuous.[10] It can be realized as a mixture of a discrete random variable and a continuous random variable; in which case the CDF will be the weighted average of the CDFs of the component variables.[10]

An example of a random variable of mixed type would be based on an experiment where a coin is flipped and the spinner is spun only if the result of the coin toss is heads. If the result is tails, X = −1; otherwise X = the value of the spinner as in the preceding example. There is a probability of 12 that this random variable will have the value −1. Other ranges of values would have half the probabilities of the last example.

Most generally, every probability distribution on the real line is a mixture of discrete part, singular part, and an absolutely continuous part; see Lebesgue's decomposition theorem § Refinement. The discrete part is concentrated on a countable set, but this set may be dense (like the set of all rational numbers).

Measure-theoretic definition edit

The most formal, axiomatic definition of a random variable involves measure theory. Continuous random variables are defined in terms of sets of numbers, along with functions that map such sets to probabilities. Because of various difficulties (e.g. the Banach–Tarski paradox) that arise if such sets are insufficiently constrained, it is necessary to introduce what is termed a sigma-algebra to constrain the possible sets over which probabilities can be defined. Normally, a particular such sigma-algebra is used, the Borel σ-algebra, which allows for probabilities to be defined over any sets that can be derived either directly from continuous intervals of numbers or by a finite or countably infinite number of unions and/or intersections of such intervals.[11]

The measure-theoretic definition is as follows.

Let   be a probability space and   a measurable space. Then an  -valued random variable is a measurable function  , which means that, for every subset  , its preimage is  -measurable;  , where  .[12] This definition enables us to measure any subset   in the target space by looking at its preimage, which by assumption is measurable.

In more intuitive terms, a member of   is a possible outcome, a member of   is a measurable subset of possible outcomes, the function   gives the probability of each such measurable subset,   represents the set of values that the random variable can take (such as the set of real numbers), and a member of   is a "well-behaved" (measurable) subset of   (those for which the probability may be determined). The random variable is then a function from any outcome to a quantity, such that the outcomes leading to any useful subset of quantities for the random variable have a well-defined probability.

When   is a topological space, then the most common choice for the σ-algebra   is the Borel σ-algebra  , which is the σ-algebra generated by the collection of all open sets in  . In such case the  -valued random variable is called an  -valued random variable. Moreover, when the space   is the real line  , then such a real-valued random variable is called simply a random variable.

Real-valued random variables edit

In this case the observation space is the set of real numbers. Recall,   is the probability space. For a real observation space, the function   is a real-valued random variable if

 

This definition is a special case of the above because the set   generates the Borel σ-algebra on the set of real numbers, and it suffices to check measurability on any generating set. Here we can prove measurability on this generating set by using the fact that  .

Moments edit

The probability distribution of a random variable is often characterised by a small number of parameters, which also have a practical interpretation. For example, it is often enough to know what its "average value" is. This is captured by the mathematical concept of expected value of a random variable, denoted  , and also called the first moment. In general,   is not equal to  . Once the "average value" is known, one could then ask how far from this average value the values of   typically are, a question that is answered by the variance and standard deviation of a random variable.   can be viewed intuitively as an average obtained from an infinite population, the members of which are particular evaluations of  .

Mathematically, this is known as the (generalised) problem of moments: for a given class of random variables  , find a collection   of functions such that the expectation values   fully characterise the distribution of the random variable  .

Moments can only be defined for real-valued functions of random variables (or complex-valued, etc.). If the random variable is itself real-valued, then moments of the variable itself can be taken, which are equivalent to moments of the identity function   of the random variable. However, even for non-real-valued random variables, moments can be taken of real-valued functions of those variables. For example, for a categorical random variable X that can take on the nominal values "red", "blue" or "green", the real-valued function   can be constructed; this uses the Iverson bracket, and has the value 1 if   has the value "green", 0 otherwise. Then, the expected value and other moments of this function can be determined.

Functions of random variables edit

A new random variable Y can be defined by applying a real Borel measurable function   to the outcomes of a real-valued random variable  . That is,  . The cumulative distribution function of   is then

 

If function   is invertible (i.e.,   exists, where   is  's inverse function) and is either increasing or decreasing, then the previous relation can be extended to obtain

 

With the same hypotheses of invertibility of  , assuming also differentiability, the relation between the probability density functions can be found by differentiating both sides of the above expression with respect to  , in order to obtain[10]

 

If there is no invertibility of   but each   admits at most a countable number of roots (i.e., a finite, or countably infinite, number of   such that  ) then the previous relation between the probability density functions can be generalized with

 

where  , according to the inverse function theorem. The formulas for densities do not demand   to be increasing.

In the measure-theoretic, axiomatic approach to probability, if a random variable   on   and a Borel measurable function  , then   is also a random variable on  , since the composition of measurable functions is also measurable. (However, this is not necessarily true if   is Lebesgue measurable.[citation needed]) The same procedure that allowed one to go from a probability space   to   can be used to obtain the distribution of  .

Example 1 edit

Let   be a real-valued, continuous random variable and let  .

 

If  , then  , so

 

If  , then

 

so

 

Example 2 edit

Suppose   is a random variable with a cumulative distribution

 

where   is a fixed parameter. Consider the random variable   Then,

 

The last expression can be calculated in terms of the cumulative distribution of   so

 

which is the cumulative distribution function (CDF) of an exponential distribution.

Example 3 edit

Suppose   is a random variable with a standard normal distribution, whose density is

 

Consider the random variable   We can find the density using the above formula for a change of variables:

 

In this case the change is not monotonic, because every value of   has two corresponding values of   (one positive and negative). However, because of symmetry, both halves will transform identically, i.e.,

 

The inverse transformation is

 

and its derivative is

 

Then,

 

This is a chi-squared distribution with one degree of freedom.

Example 4 edit

Suppose   is a random variable with a normal distribution, whose density is

 

Consider the random variable   We can find the density using the above formula for a change of variables:

 

In this case the change is not monotonic, because every value of   has two corresponding values of   (one positive and negative). Differently from the previous example, in this case however, there is no symmetry and we have to compute the two distinct terms:

 

The inverse transformation is

 

and its derivative is

 

Then,

 

This is a noncentral chi-squared distribution with one degree of freedom.

Some properties edit

  • The probability distribution of the sum of two independent random variables is the convolution of each of their distributions.
  • Probability distributions are not a vector space—they are not closed under linear combinations, as these do not preserve non-negativity or total integral 1—but they are closed under convex combination, thus forming a convex subset of the space of functions (or measures).

Equivalence of random variables edit

There are several different senses in which random variables can be considered to be equivalent. Two random variables can be equal, equal almost surely, or equal in distribution.

In increasing order of strength, the precise definition of these notions of equivalence is given below.

Equality in distribution edit

If the sample space is a subset of the real line, random variables X and Y are equal in distribution (denoted  ) if they have the same distribution functions:

 

To be equal in distribution, random variables need not be defined on the same probability space. Two random variables having equal moment generating functions have the same distribution. This provides, for example, a useful method of checking equality of certain functions of independent, identically distributed (IID) random variables. However, the moment generating function exists only for distributions that have a defined Laplace transform.

Almost sure equality edit

Two random variables X and Y are equal almost surely (denoted  ) if, and only if, the probability that they are different is zero:

 

For all practical purposes in probability theory, this notion of equivalence is as strong as actual equality. It is associated to the following distance:

 

where "ess sup" represents the essential supremum in the sense of measure theory.

Equality edit

Finally, the two random variables X and Y are equal if they are equal as functions on their measurable space:

 

This notion is typically the least useful in probability theory because in practice and in theory, the underlying measure space of the experiment is rarely explicitly characterized or even characterizable.

Convergence edit

A significant theme in mathematical statistics consists of obtaining convergence results for certain sequences of random variables; for instance the law of large numbers and the central limit theorem.

There are various senses in which a sequence   of random variables can converge to a random variable  . These are explained in the article on convergence of random variables.

See also edit

References edit

Inline citations edit

  1. ^ a b Blitzstein, Joe; Hwang, Jessica (2014). Introduction to Probability. CRC Press. ISBN 9781466575592.
  2. ^ Deisenroth, Marc Peter (2020). Mathematics for machine learning. A. Aldo Faisal, Cheng Soon Ong. Cambridge, United Kingdom: Cambridge University Press. ISBN 978-1-108-47004-9. OCLC 1104219401.
  3. ^ George Mackey (July 1980). "Harmonic analysis as the exploitation of symmetry – a historical survey". Bulletin of the American Mathematical Society. New Series. 3 (1).
  4. ^ "Random Variables". www.mathsisfun.com. Retrieved 2020-08-21.
  5. ^ Yates, Daniel S.; Moore, David S; Starnes, Daren S. (2003). (2nd ed.). New York: Freeman. ISBN 978-0-7167-4773-4. Archived from the original on 2005-02-09.
  6. ^ "Random Variables". www.stat.yale.edu. Retrieved 2020-08-21.
  7. ^ Dekking, Frederik Michel; Kraaikamp, Cornelis; Lopuhaä, Hendrik Paul; Meester, Ludolf Erwin (2005). "A Modern Introduction to Probability and Statistics". Springer Texts in Statistics. doi:10.1007/1-84628-168-7. ISBN 978-1-85233-896-1. ISSN 1431-875X.
  8. ^ L. Castañeda; V. Arunachalam & S. Dharmaraja (2012). Introduction to Probability and Stochastic Processes with Applications. Wiley. p. 67. ISBN 9781118344941.
  9. ^ Billingsley, Patrick (1995). Probability and Measure (3rd ed.). Wiley. p. 187. ISBN 9781466575592.
  10. ^ a b c d Bertsekas, Dimitri P. (2002). Introduction to Probability. Tsitsiklis, John N., Τσιτσικλής, Γιάννης Ν. Belmont, Mass.: Athena Scientific. ISBN 188652940X. OCLC 51441829.
  11. ^ Steigerwald, Douglas G. "Economics 245A – Introduction to Measure Theory" (PDF). University of California, Santa Barbara. Retrieved April 26, 2013.
  12. ^ Fristedt & Gray (1996, page 11)

Literature edit

External links edit

  • "Random variable", Encyclopedia of Mathematics, EMS Press, 2001 [1994]
  • Zukerman, Moshe (2014), Introduction to Queueing Theory and Stochastic Teletraffic Models (PDF), arXiv:1307.2968
  • Zukerman, Moshe (2014), Basic Probability Topics (PDF)

random, variable, random, variable, also, called, random, quantity, aleatory, variable, stochastic, variable, mathematical, formalization, quantity, object, which, depends, random, events, term, random, variable, mathematical, definition, refers, neither, rand. A random variable also called random quantity aleatory variable or stochastic variable is a mathematical formalization of a quantity or object which depends on random events 1 The term random variable in its mathematical definition refers to neither randomness nor variability 2 but instead is a mathematical function in which the domain is the set of possible outcomes in a sample space e g the set H T displaystyle H T which are the possible upper sides of a flipped coin heads H displaystyle H or tails T displaystyle T as the result from tossing a coin and the range is a measurable space e g corresponding to the domain above the range might be the set 1 1 displaystyle 1 1 if say heads H displaystyle H mapped to 1 and T displaystyle T mapped to 1 Typically the range of a random variable is set of real numbers This graph shows how random variable is a function from all possible outcomes to real values It also shows how random variable is used for defining probability mass functions Informally randomness typically represents some fundamental element of chance such as in the roll of a die it may also represent uncertainty such as measurement error 1 However the interpretation of probability is philosophically complicated and even in specific cases is not always straightforward The purely mathematical analysis of random variables is independent of such interpretational difficulties and can be based upon a rigorous axiomatic setup In the formal mathematical language of measure theory a random variable is defined as a measurable function from a probability measure space called the sample space to a measurable space This allows consideration of the pushforward measure which is called the distribution of the random variable the distribution is thus a probability measure on the set of all possible values of the random variable It is possible for two random variables to have identical distributions but to differ in significant ways for instance they may be independent It is common to consider the special cases of discrete random variables and absolutely continuous random variables corresponding to whether a random variable is valued in a countable subset or in an interval of real numbers There are other important possibilities especially in the theory of stochastic processes wherein it is natural to consider random sequences or random functions Sometimes a random variable is taken to be automatically valued in the real numbers with more general random quantities instead being called random elements According to George Mackey Pafnuty Chebyshev was the first person to think systematically in terms of random variables 3 Contents 1 Definition 1 1 Standard case 1 2 Extensions 2 Distribution functions 3 Examples 3 1 Discrete random variable 3 1 1 Coin toss 3 1 2 Dice roll 3 2 Continuous random variable 3 3 Mixed type 4 Measure theoretic definition 4 1 Real valued random variables 5 Moments 6 Functions of random variables 6 1 Example 1 6 2 Example 2 6 3 Example 3 6 4 Example 4 7 Some properties 8 Equivalence of random variables 8 1 Equality in distribution 8 2 Almost sure equality 8 3 Equality 9 Convergence 10 See also 11 References 11 1 Inline citations 11 2 Literature 12 External linksDefinition editA random variable X displaystyle X nbsp is a measurable function X W E displaystyle X colon Omega to E nbsp from a sample space W displaystyle Omega nbsp as a set of possible outcomes to a measurable space E displaystyle E nbsp The technical axiomatic definition requires the sample space W displaystyle Omega nbsp to be a sample space of a probability triple W F P displaystyle Omega mathcal F operatorname P nbsp see the measure theoretic definition A random variable is often denoted by capital Roman letters such as X Y Z T displaystyle X Y Z T nbsp 4 The probability that X displaystyle X nbsp takes on a value in a measurable set S E displaystyle S subseteq E nbsp is written as P X S P w W X w S displaystyle operatorname P X in S operatorname P omega in Omega mid X omega in S nbsp Standard case edit In many cases X displaystyle X nbsp is real valued i e E R displaystyle E mathbb R nbsp In some contexts the term random element see extensions is used to denote a random variable not of this form When the image or range of X displaystyle X nbsp is finitely or infinitely countable the random variable is called a discrete random variable 5 399 and its distribution is a discrete probability distribution i e can be described by a probability mass function that assigns a probability to each value in the image of X displaystyle X nbsp If the image is uncountably infinite usually an interval then X displaystyle X nbsp is called a continuous random variable 6 7 In the special case that it is absolutely continuous its distribution can be described by a probability density function which assigns probabilities to intervals in particular each individual point must necessarily have probability zero for an absolutely continuous random variable Not all continuous random variables are absolutely continuous 8 Any random variable can be described by its cumulative distribution function which describes the probability that the random variable will be less than or equal to a certain value Extensions edit The term random variable in statistics is traditionally limited to the real valued case E R displaystyle E mathbb R nbsp In this case the structure of the real numbers makes it possible to define quantities such as the expected value and variance of a random variable its cumulative distribution function and the moments of its distribution However the definition above is valid for any measurable space E displaystyle E nbsp of values Thus one can consider random elements of other sets E displaystyle E nbsp such as random Boolean values categorical values complex numbers vectors matrices sequences trees sets shapes manifolds and functions One may then specifically refer to a random variable of type E displaystyle E nbsp or an E displaystyle E nbsp valued random variable This more general concept of a random element is particularly useful in disciplines such as graph theory machine learning natural language processing and other fields in discrete mathematics and computer science where one is often interested in modeling the random variation of non numerical data structures In some cases it is nonetheless convenient to represent each element of E displaystyle E nbsp using one or more real numbers In this case a random element may optionally be represented as a vector of real valued random variables all defined on the same underlying probability space W displaystyle Omega nbsp which allows the different random variables to covary For example A random word may be represented as a random integer that serves as an index into the vocabulary of possible words Alternatively it can be represented as a random indicator vector whose length equals the size of the vocabulary where the only values of positive probability are 1 0 0 0 displaystyle 1 0 0 0 cdots nbsp 0 1 0 0 displaystyle 0 1 0 0 cdots nbsp 0 0 1 0 displaystyle 0 0 1 0 cdots nbsp and the position of the 1 indicates the word A random sentence of given length N displaystyle N nbsp may be represented as a vector of N displaystyle N nbsp random words A random graph on N displaystyle N nbsp given vertices may be represented as a N N displaystyle N times N nbsp matrix of random variables whose values specify the adjacency matrix of the random graph A random function F displaystyle F nbsp may be represented as a collection of random variables F x displaystyle F x nbsp giving the function s values at the various points x displaystyle x nbsp in the function s domain The F x displaystyle F x nbsp are ordinary real valued random variables provided that the function is real valued For example a stochastic process is a random function of time a random vector is a random function of some index set such as 1 2 n displaystyle 1 2 ldots n nbsp and random field is a random function on any set typically time space or a discrete set Distribution functions editIf a random variable X W R displaystyle X colon Omega to mathbb R nbsp defined on the probability space W F P displaystyle Omega mathcal F operatorname P nbsp is given we can ask questions like How likely is it that the value of X displaystyle X nbsp is equal to 2 This is the same as the probability of the event w X w 2 displaystyle omega X omega 2 nbsp which is often written as P X 2 displaystyle P X 2 nbsp or p X 2 displaystyle p X 2 nbsp for short Recording all these probabilities of outputs of a random variable X displaystyle X nbsp yields the probability distribution of X displaystyle X nbsp The probability distribution forgets about the particular probability space used to define X displaystyle X nbsp and only records the probabilities of various output values of X displaystyle X nbsp Such a probability distribution if X displaystyle X nbsp is real valued can always be captured by its cumulative distribution function F X x P X x displaystyle F X x operatorname P X leq x nbsp and sometimes also using a probability density function f X displaystyle f X nbsp In measure theoretic terms we use the random variable X displaystyle X nbsp to push forward the measure P displaystyle P nbsp on W displaystyle Omega nbsp to a measure p X displaystyle p X nbsp on R displaystyle mathbb R nbsp The measure p X displaystyle p X nbsp is called the probability distribution of X displaystyle X nbsp or the law of X displaystyle X nbsp 9 The density f X d p X d m displaystyle f X dp X d mu nbsp the Radon Nikodym derivative of p X displaystyle p X nbsp with respect to some reference measure m displaystyle mu nbsp on R displaystyle mathbb R nbsp often this reference measure is the Lebesgue measure in the case of continuous random variables or the counting measure in the case of discrete random variables The underlying probability space W displaystyle Omega nbsp is a technical device used to guarantee the existence of random variables sometimes to construct them and to define notions such as correlation and dependence or independence based on a joint distribution of two or more random variables on the same probability space In practice one often disposes of the space W displaystyle Omega nbsp altogether and just puts a measure on R displaystyle mathbb R nbsp that assigns measure 1 to the whole real line i e one works with probability distributions instead of random variables See the article on quantile functions for fuller development Examples editDiscrete random variable edit Consider an experiment where a person is chosen at random An example of a random variable may be the person s height Mathematically the random variable is interpreted as a function which maps the person to his or her height Associated with the random variable is a probability distribution that allows the computation of the probability that the height is in any subset of possible values such as the probability that the height is between 180 and 190 cm or the probability that the height is either less than 150 or more than 200 cm Another random variable may be the person s number of children this is a discrete random variable with non negative integer values It allows the computation of probabilities for individual integer values the probability mass function PMF or for sets of values including infinite sets For example the event of interest may be an even number of children For both finite and infinite event sets their probabilities can be found by adding up the PMFs of the elements that is the probability of an even number of children is the infinite sum PMF 0 PMF 2 PMF 4 displaystyle operatorname PMF 0 operatorname PMF 2 operatorname PMF 4 cdots nbsp In examples such as these the sample space is often suppressed since it is mathematically hard to describe and the possible values of the random variables are then treated as a sample space But when two random variables are measured on the same sample space of outcomes such as the height and number of children being computed on the same random persons it is easier to track their relationship if it is acknowledged that both height and number of children come from the same random person for example so that questions of whether such random variables are correlated or not can be posed If a n b n textstyle a n b n nbsp are countable sets of real numbers b n gt 0 textstyle b n gt 0 nbsp and n b n 1 textstyle sum n b n 1 nbsp then F n b n d a n x textstyle F sum n b n delta a n x nbsp is a discrete distribution function Here d t x 0 displaystyle delta t x 0 nbsp for x lt t displaystyle x lt t nbsp d t x 1 displaystyle delta t x 1 nbsp for x t displaystyle x geq t nbsp Taking for instance an enumeration of all rational numbers as a n displaystyle a n nbsp one gets a discrete function that is not necessarily a step function piecewise constant Coin toss edit The possible outcomes for one coin toss can be described by the sample space W heads tails displaystyle Omega text heads text tails nbsp We can introduce a real valued random variable Y displaystyle Y nbsp that models a 1 payoff for a successful bet on heads as follows Y w 1 if w heads 0 if w tails displaystyle Y omega begin cases 1 amp text if omega text heads 6pt 0 amp text if omega text tails end cases nbsp If the coin is a fair coin Y has a probability mass function f Y displaystyle f Y nbsp given by f Y y 1 2 if y 1 1 2 if y 0 displaystyle f Y y begin cases tfrac 1 2 amp text if y 1 6pt tfrac 1 2 amp text if y 0 end cases nbsp Dice roll edit nbsp If the sample space is the set of possible numbers rolled on two dice and the random variable of interest is the sum S of the numbers on the two dice then S is a discrete random variable whose distribution is described by the probability mass function plotted as the height of picture columns here A random variable can also be used to describe the process of rolling dice and the possible outcomes The most obvious representation for the two dice case is to take the set of pairs of numbers n1 and n2 from 1 2 3 4 5 6 representing the numbers on the two dice as the sample space The total number rolled the sum of the numbers in each pair is then a random variable X given by the function that maps the pair to the sum X n 1 n 2 n 1 n 2 displaystyle X n 1 n 2 n 1 n 2 nbsp and if the dice are fair has a probability mass function fX given by f X S min S 1 13 S 36 for S 2 3 4 5 6 7 8 9 10 11 12 displaystyle f X S frac min S 1 13 S 36 text for S in 2 3 4 5 6 7 8 9 10 11 12 nbsp Continuous random variable edit Formally a continuous random variable is a random variable whose cumulative distribution function is continuous everywhere 10 There are no gaps which would correspond to numbers which have a finite probability of occurring Instead continuous random variables almost never take an exact prescribed value c formally c R Pr X c 0 textstyle forall c in mathbb R Pr X c 0 nbsp but there is a positive probability that its value will lie in particular intervals which can be arbitrarily small Continuous random variables usually admit probability density functions PDF which characterize their CDF and probability measures such distributions are also called absolutely continuous but some continuous distributions are singular or mixes of an absolutely continuous part and a singular part An example of a continuous random variable would be one based on a spinner that can choose a horizontal direction Then the values taken by the random variable are directions We could represent these directions by North West East South Southeast etc However it is commonly more convenient to map the sample space to a random variable which takes values which are real numbers This can be done for example by mapping a direction to a bearing in degrees clockwise from North The random variable then takes values which are real numbers from the interval 0 360 with all parts of the range being equally likely In this case X the angle spun Any real number has probability zero of being selected but a positive probability can be assigned to any range of values For example the probability of choosing a number in 0 180 is 1 2 Instead of speaking of a probability mass function we say that the probability density of X is 1 360 The probability of a subset of 0 360 can be calculated by multiplying the measure of the set by 1 360 In general the probability of a set for a given continuous random variable can be calculated by integrating the density over the given set More formally given any interval I a b x R a x b textstyle I a b x in mathbb R a leq x leq b nbsp a random variable X I U I U a b displaystyle X I sim operatorname U I operatorname U a b nbsp is called a continuous uniform random variable CURV if the probability that it takes a value in a subinterval depends only on the length of the subinterval This implies that the probability of X I displaystyle X I nbsp falling in any subinterval c d a b displaystyle c d subseteq a b nbsp is proportional to the length of the subinterval that is if a c d b one hasPr X I c d d c b a displaystyle Pr left X I in c d right frac d c b a nbsp where the last equality results from the unitarity axiom of probability The probability density function of a CURV X U a b displaystyle X sim operatorname U a b nbsp is given by the indicator function of its interval of support normalized by the interval s length f X x 1 b a a x b 0 otherwise displaystyle f X x begin cases displaystyle 1 over b a amp a leq x leq b 0 amp text otherwise end cases nbsp Of particular interest is the uniform distribution on the unit interval 0 1 displaystyle 0 1 nbsp Samples of any desired probability distribution D displaystyle operatorname D nbsp can be generated by calculating the quantile function of D displaystyle operatorname D nbsp on a randomly generated number distributed uniformly on the unit interval This exploits properties of cumulative distribution functions which are a unifying framework for all random variables Mixed type edit A mixed random variable is a random variable whose cumulative distribution function is neither discrete nor everywhere continuous 10 It can be realized as a mixture of a discrete random variable and a continuous random variable in which case the CDF will be the weighted average of the CDFs of the component variables 10 An example of a random variable of mixed type would be based on an experiment where a coin is flipped and the spinner is spun only if the result of the coin toss is heads If the result is tails X 1 otherwise X the value of the spinner as in the preceding example There is a probability of 1 2 that this random variable will have the value 1 Other ranges of values would have half the probabilities of the last example Most generally every probability distribution on the real line is a mixture of discrete part singular part and an absolutely continuous part see Lebesgue s decomposition theorem Refinement The discrete part is concentrated on a countable set but this set may be dense like the set of all rational numbers Measure theoretic definition editThe most formal axiomatic definition of a random variable involves measure theory Continuous random variables are defined in terms of sets of numbers along with functions that map such sets to probabilities Because of various difficulties e g the Banach Tarski paradox that arise if such sets are insufficiently constrained it is necessary to introduce what is termed a sigma algebra to constrain the possible sets over which probabilities can be defined Normally a particular such sigma algebra is used the Borel s algebra which allows for probabilities to be defined over any sets that can be derived either directly from continuous intervals of numbers or by a finite or countably infinite number of unions and or intersections of such intervals 11 The measure theoretic definition is as follows Let W F P displaystyle Omega mathcal F P nbsp be a probability space and E E displaystyle E mathcal E nbsp a measurable space Then an E E displaystyle E mathcal E nbsp valued random variable is a measurable function X W E displaystyle X colon Omega to E nbsp which means that for every subset B E displaystyle B in mathcal E nbsp its preimage is F displaystyle mathcal F nbsp measurable X 1 B F displaystyle X 1 B in mathcal F nbsp where X 1 B w X w B displaystyle X 1 B omega X omega in B nbsp 12 This definition enables us to measure any subset B E displaystyle B in mathcal E nbsp in the target space by looking at its preimage which by assumption is measurable In more intuitive terms a member of W displaystyle Omega nbsp is a possible outcome a member of F displaystyle mathcal F nbsp is a measurable subset of possible outcomes the function P displaystyle P nbsp gives the probability of each such measurable subset E displaystyle E nbsp represents the set of values that the random variable can take such as the set of real numbers and a member of E displaystyle mathcal E nbsp is a well behaved measurable subset of E displaystyle E nbsp those for which the probability may be determined The random variable is then a function from any outcome to a quantity such that the outcomes leading to any useful subset of quantities for the random variable have a well defined probability When E displaystyle E nbsp is a topological space then the most common choice for the s algebra E displaystyle mathcal E nbsp is the Borel s algebra B E displaystyle mathcal B E nbsp which is the s algebra generated by the collection of all open sets in E displaystyle E nbsp In such case the E E displaystyle E mathcal E nbsp valued random variable is called an E displaystyle E nbsp valued random variable Moreover when the space E displaystyle E nbsp is the real line R displaystyle mathbb R nbsp then such a real valued random variable is called simply a random variable Real valued random variables edit In this case the observation space is the set of real numbers Recall W F P displaystyle Omega mathcal F P nbsp is the probability space For a real observation space the function X W R displaystyle X colon Omega rightarrow mathbb R nbsp is a real valued random variable if w X w r F r R displaystyle omega X omega leq r in mathcal F qquad forall r in mathbb R nbsp This definition is a special case of the above because the set r r R displaystyle infty r r in mathbb R nbsp generates the Borel s algebra on the set of real numbers and it suffices to check measurability on any generating set Here we can prove measurability on this generating set by using the fact that w X w r X 1 r displaystyle omega X omega leq r X 1 infty r nbsp Moments editThe probability distribution of a random variable is often characterised by a small number of parameters which also have a practical interpretation For example it is often enough to know what its average value is This is captured by the mathematical concept of expected value of a random variable denoted E X displaystyle operatorname E X nbsp and also called the first moment In general E f X displaystyle operatorname E f X nbsp is not equal to f E X displaystyle f operatorname E X nbsp Once the average value is known one could then ask how far from this average value the values of X displaystyle X nbsp typically are a question that is answered by the variance and standard deviation of a random variable E X displaystyle operatorname E X nbsp can be viewed intuitively as an average obtained from an infinite population the members of which are particular evaluations of X displaystyle X nbsp Mathematically this is known as the generalised problem of moments for a given class of random variables X displaystyle X nbsp find a collection f i displaystyle f i nbsp of functions such that the expectation values E f i X displaystyle operatorname E f i X nbsp fully characterise the distribution of the random variable X displaystyle X nbsp Moments can only be defined for real valued functions of random variables or complex valued etc If the random variable is itself real valued then moments of the variable itself can be taken which are equivalent to moments of the identity function f X X displaystyle f X X nbsp of the random variable However even for non real valued random variables moments can be taken of real valued functions of those variables For example for a categorical random variable X that can take on the nominal values red blue or green the real valued function X green displaystyle X text green nbsp can be constructed this uses the Iverson bracket and has the value 1 if X displaystyle X nbsp has the value green 0 otherwise Then the expected value and other moments of this function can be determined Functions of random variables editA new random variable Y can be defined by applying a real Borel measurable function g R R displaystyle g colon mathbb R rightarrow mathbb R nbsp to the outcomes of a real valued random variable X displaystyle X nbsp That is Y g X displaystyle Y g X nbsp The cumulative distribution function of Y displaystyle Y nbsp is then F Y y P g X y displaystyle F Y y operatorname P g X leq y nbsp If function g displaystyle g nbsp is invertible i e h g 1 displaystyle h g 1 nbsp exists where h displaystyle h nbsp is g displaystyle g nbsp s inverse function and is either increasing or decreasing then the previous relation can be extended to obtain F Y y P g X y P X h y F X h y if h g 1 increasing P X h y 1 F X h y if h g 1 decreasing displaystyle F Y y operatorname P g X leq y begin cases operatorname P X leq h y F X h y amp text if h g 1 text increasing operatorname P X geq h y 1 F X h y amp text if h g 1 text decreasing end cases nbsp With the same hypotheses of invertibility of g displaystyle g nbsp assuming also differentiability the relation between the probability density functions can be found by differentiating both sides of the above expression with respect to y displaystyle y nbsp in order to obtain 10 f Y y f X h y d h y d y displaystyle f Y y f X bigl h y bigr left frac dh y dy right nbsp If there is no invertibility of g displaystyle g nbsp but each y displaystyle y nbsp admits at most a countable number of roots i e a finite or countably infinite number of x i displaystyle x i nbsp such that y g x i displaystyle y g x i nbsp then the previous relation between the probability density functions can be generalized with f Y y i f X g i 1 y d g i 1 y d y displaystyle f Y y sum i f X g i 1 y left frac dg i 1 y dy right nbsp where x i g i 1 y displaystyle x i g i 1 y nbsp according to the inverse function theorem The formulas for densities do not demand g displaystyle g nbsp to be increasing In the measure theoretic axiomatic approach to probability if a random variable X displaystyle X nbsp on W displaystyle Omega nbsp and a Borel measurable function g R R displaystyle g colon mathbb R rightarrow mathbb R nbsp then Y g X displaystyle Y g X nbsp is also a random variable on W displaystyle Omega nbsp since the composition of measurable functions is also measurable However this is not necessarily true if g displaystyle g nbsp is Lebesgue measurable citation needed The same procedure that allowed one to go from a probability space W P displaystyle Omega P nbsp to R d F X displaystyle mathbb R dF X nbsp can be used to obtain the distribution of Y displaystyle Y nbsp Example 1 edit Let X displaystyle X nbsp be a real valued continuous random variable and let Y X 2 displaystyle Y X 2 nbsp F Y y P X 2 y displaystyle F Y y operatorname P X 2 leq y nbsp If y lt 0 displaystyle y lt 0 nbsp then P X 2 y 0 displaystyle P X 2 leq y 0 nbsp so F Y y 0 if y lt 0 displaystyle F Y y 0 qquad hbox if quad y lt 0 nbsp If y 0 displaystyle y geq 0 nbsp then P X 2 y P X y P y X y displaystyle operatorname P X 2 leq y operatorname P X leq sqrt y operatorname P sqrt y leq X leq sqrt y nbsp so F Y y F X y F X y if y 0 displaystyle F Y y F X sqrt y F X sqrt y qquad hbox if quad y geq 0 nbsp Example 2 edit Suppose X displaystyle X nbsp is a random variable with a cumulative distribution F X x P X x 1 1 e x 8 displaystyle F X x P X leq x frac 1 1 e x theta nbsp where 8 gt 0 displaystyle theta gt 0 nbsp is a fixed parameter Consider the random variable Y l o g 1 e X displaystyle Y mathrm log 1 e X nbsp Then F Y y P Y y P l o g 1 e X y P X l o g e y 1 displaystyle F Y y P Y leq y P mathrm log 1 e X leq y P X geq mathrm log e y 1 nbsp The last expression can be calculated in terms of the cumulative distribution of X displaystyle X nbsp so F Y y 1 F X log e y 1 1 1 1 e log e y 1 8 1 1 1 e y 1 8 1 e y 8 displaystyle begin aligned F Y y amp 1 F X log e y 1 5pt amp 1 frac 1 1 e log e y 1 theta 5pt amp 1 frac 1 1 e y 1 theta 5pt amp 1 e y theta end aligned nbsp which is the cumulative distribution function CDF of an exponential distribution Example 3 edit Suppose X displaystyle X nbsp is a random variable with a standard normal distribution whose density is f X x 1 2 p e x 2 2 displaystyle f X x frac 1 sqrt 2 pi e x 2 2 nbsp Consider the random variable Y X 2 displaystyle Y X 2 nbsp We can find the density using the above formula for a change of variables f Y y i f X g i 1 y d g i 1 y d y displaystyle f Y y sum i f X g i 1 y left frac dg i 1 y dy right nbsp In this case the change is not monotonic because every value of Y displaystyle Y nbsp has two corresponding values of X displaystyle X nbsp one positive and negative However because of symmetry both halves will transform identically i e f Y y 2 f X g 1 y d g 1 y d y displaystyle f Y y 2f X g 1 y left frac dg 1 y dy right nbsp The inverse transformation is x g 1 y y displaystyle x g 1 y sqrt y nbsp and its derivative is d g 1 y d y 1 2 y displaystyle frac dg 1 y dy frac 1 2 sqrt y nbsp Then f Y y 2 1 2 p e y 2 1 2 y 1 2 p y e y 2 displaystyle f Y y 2 frac 1 sqrt 2 pi e y 2 frac 1 2 sqrt y frac 1 sqrt 2 pi y e y 2 nbsp This is a chi squared distribution with one degree of freedom Example 4 edit Suppose X displaystyle X nbsp is a random variable with a normal distribution whose density is f X x 1 2 p s 2 e x m 2 2 s 2 displaystyle f X x frac 1 sqrt 2 pi sigma 2 e x mu 2 2 sigma 2 nbsp Consider the random variable Y X 2 displaystyle Y X 2 nbsp We can find the density using the above formula for a change of variables f Y y i f X g i 1 y d g i 1 y d y displaystyle f Y y sum i f X g i 1 y left frac dg i 1 y dy right nbsp In this case the change is not monotonic because every value of Y displaystyle Y nbsp has two corresponding values of X displaystyle X nbsp one positive and negative Differently from the previous example in this case however there is no symmetry and we have to compute the two distinct terms f Y y f X g 1 1 y d g 1 1 y d y f X g 2 1 y d g 2 1 y d y displaystyle f Y y f X g 1 1 y left frac dg 1 1 y dy right f X g 2 1 y left frac dg 2 1 y dy right nbsp The inverse transformation is x g 1 2 1 y y displaystyle x g 1 2 1 y pm sqrt y nbsp and its derivative is d g 1 2 1 y d y 1 2 y displaystyle frac dg 1 2 1 y dy pm frac 1 2 sqrt y nbsp Then f Y y 1 2 p s 2 1 2 y e y m 2 2 s 2 e y m 2 2 s 2 displaystyle f Y y frac 1 sqrt 2 pi sigma 2 frac 1 2 sqrt y e sqrt y mu 2 2 sigma 2 e sqrt y mu 2 2 sigma 2 nbsp This is a noncentral chi squared distribution with one degree of freedom Some properties editThe probability distribution of the sum of two independent random variables is the convolution of each of their distributions Probability distributions are not a vector space they are not closed under linear combinations as these do not preserve non negativity or total integral 1 but they are closed under convex combination thus forming a convex subset of the space of functions or measures Equivalence of random variables editThere are several different senses in which random variables can be considered to be equivalent Two random variables can be equal equal almost surely or equal in distribution In increasing order of strength the precise definition of these notions of equivalence is given below Equality in distribution edit If the sample space is a subset of the real line random variables X and Y are equal in distribution denoted X d Y displaystyle X stackrel d Y nbsp if they have the same distribution functions P X x P Y x for all x displaystyle operatorname P X leq x operatorname P Y leq x quad text for all x nbsp To be equal in distribution random variables need not be defined on the same probability space Two random variables having equal moment generating functions have the same distribution This provides for example a useful method of checking equality of certain functions of independent identically distributed IID random variables However the moment generating function exists only for distributions that have a defined Laplace transform Almost sure equality edit Two random variables X and Y are equal almost surely denoted X a s Y displaystyle X stackrel text a s Y nbsp if and only if the probability that they are different is zero P X Y 0 displaystyle operatorname P X neq Y 0 nbsp For all practical purposes in probability theory this notion of equivalence is as strong as actual equality It is associated to the following distance d X Y ess sup w X w Y w displaystyle d infty X Y operatorname ess sup omega X omega Y omega nbsp where ess sup represents the essential supremum in the sense of measure theory Equality edit Finally the two random variables X and Y are equal if they are equal as functions on their measurable space X w Y w for all w displaystyle X omega Y omega qquad hbox for all omega nbsp This notion is typically the least useful in probability theory because in practice and in theory the underlying measure space of the experiment is rarely explicitly characterized or even characterizable Convergence editMain article Convergence of random variables A significant theme in mathematical statistics consists of obtaining convergence results for certain sequences of random variables for instance the law of large numbers and the central limit theorem There are various senses in which a sequence X n displaystyle X n nbsp of random variables can converge to a random variable X displaystyle X nbsp These are explained in the article on convergence of random variables See also edit nbsp Mathematics portal Aleatoricism Algebra of random variables Event probability theory Multivariate random variable Pairwise independent random variables Observable variable Random compact set Random element Random function Random measure Random number generator Random variate Random vector Randomness Stochastic process Relationships among probability distributionsReferences editInline citations edit a b Blitzstein Joe Hwang Jessica 2014 Introduction to Probability CRC Press ISBN 9781466575592 Deisenroth Marc Peter 2020 Mathematics for machine learning A Aldo Faisal Cheng Soon Ong Cambridge United Kingdom Cambridge University Press ISBN 978 1 108 47004 9 OCLC 1104219401 George Mackey July 1980 Harmonic analysis as the exploitation of symmetry a historical survey Bulletin of the American Mathematical Society New Series 3 1 Random Variables www mathsisfun com Retrieved 2020 08 21 Yates Daniel S Moore David S Starnes Daren S 2003 The Practice of Statistics 2nd ed New York Freeman ISBN 978 0 7167 4773 4 Archived from the original on 2005 02 09 Random Variables www stat yale edu Retrieved 2020 08 21 Dekking Frederik Michel Kraaikamp Cornelis Lopuhaa Hendrik Paul Meester Ludolf Erwin 2005 A Modern Introduction to Probability and Statistics Springer Texts in Statistics doi 10 1007 1 84628 168 7 ISBN 978 1 85233 896 1 ISSN 1431 875X L Castaneda V Arunachalam amp S Dharmaraja 2012 Introduction to Probability and Stochastic Processes with Applications Wiley p 67 ISBN 9781118344941 Billingsley Patrick 1995 Probability and Measure 3rd ed Wiley p 187 ISBN 9781466575592 a b c d Bertsekas Dimitri P 2002 Introduction to Probability Tsitsiklis John N Tsitsiklhs Giannhs N Belmont Mass Athena Scientific ISBN 188652940X OCLC 51441829 Steigerwald Douglas G Economics 245A Introduction to Measure Theory PDF University of California Santa Barbara Retrieved April 26 2013 Fristedt amp Gray 1996 page 11 Literature edit Fristedt Bert Gray Lawrence 1996 A modern approach to probability theory Boston Birkhauser ISBN 3 7643 3807 5 Billingsley Patrick 1995 Probability and Measure New York Wiley ISBN 8126517719 Kallenberg Olav 1986 Random Measures 4th ed Berlin Akademie Verlag ISBN 0 12 394960 2 MR 0854102 Kallenberg Olav 2001 Foundations of Modern Probability 2nd ed Berlin Springer Verlag ISBN 0 387 95313 2 Papoulis Athanasios 1965 Probability Random Variables and Stochastic Processes 9th ed Tokyo McGraw Hill ISBN 0 07 119981 0 External links edit Random variable Encyclopedia of Mathematics EMS Press 2001 1994 Zukerman Moshe 2014 Introduction to Queueing Theory and Stochastic Teletraffic Models PDF arXiv 1307 2968 Zukerman Moshe 2014 Basic Probability Topics PDF Retrieved from https en wikipedia org w index php title Random variable amp oldid 1220695751, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.