Random element

Summary

In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by Maurice Fréchet (1948) who commented that the “development of probability theory and expansion of area of its applications have led to necessity to pass from schemes where (random) outcomes of experiments can be described by number or a finite set of numbers, to schemes where outcomes of experiments represent, for example, vectors, functions, processes, fields, series, transformations, and also sets or collections of sets.”[1]

The modern-day usage of “random element” frequently assumes the space of values is a topological vector space, often a Banach or Hilbert space with a specified natural sigma algebra of subsets.[2]

Definition edit

Let   be a probability space, and   a measurable space. A random element with values in E is a function X: Ω→E which is  -measurable. That is, a function X such that for any  , the preimage of B lies in  .

Sometimes random elements with values in   are called  -valued random variables.

Note if  , where   are the real numbers, and   is its Borel σ-algebra, then the definition of random element is the classical definition of random variable.

The definition of a random element   with values in a Banach space   is typically understood to utilize the smallest  -algebra on B for which every bounded linear functional is measurable. An equivalent definition, in this case, to the above, is that a map  , from a probability space, is a random element if   is a random variable for every bounded linear functional f, or, equivalently, that   is weakly measurable.

Examples of random elements edit

Random variable edit

A random variable is the simplest type of random element. It is a map   is a measurable function from the set of possible outcomes   to  .

As a real-valued function,   often describes some numerical quantity of a given event. E.g. the number of heads after a certain number of coin flips; the heights of different people.

When the image (or range) of   is finite or countably infinite, the random variable is called a discrete random variable[3] and its distribution can be described by a probability mass function which assigns a probability to each value in the image of  . If the image is uncountably infinite then   is called a continuous random variable. In the special case that it is absolutely continuous, its distribution can be described by a probability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous,[4] for example a mixture distribution. Such random variables cannot be described by a probability density or a probability mass function.

Random vector edit

A random vector is a column vector   (or its transpose, which is a row vector) whose components are scalar-valued random variables on the same probability space  , where   is the sample space,   is the sigma-algebra (the collection of all events), and   is the probability measure (a function returning each event's probability).

Random vectors are often used as the underlying implementation of various types of aggregate random variables, e.g. a random matrix, random tree, random sequence, random process, etc.

Random matrix edit

A random matrix is a matrix-valued random element. Many important properties of physical systems can be represented mathematically as matrix problems. For example, the thermal conductivity of a lattice can be computed from the dynamical matrix of the particle-particle interactions within the lattice.

Random function edit

A random function is a type of random element in which a single outcome is selected from some family of functions, where the family consists some class of all maps from the domain to the codomain. For example, the class may be restricted to all continuous functions or to all step functions. The values determined by a random function evaluated at different points from the same realization would not generally be statistically independent but, depending on the model, values determined at the same or different points from different realisations might well be treated as independent.

Random process edit

A Random process is a collection of random variables, representing the evolution of some system of random values over time. This is the probabilistic counterpart to a deterministic process (or deterministic system). Instead of describing a process which can only evolve in one way (as in the case, for example, of solutions of an ordinary differential equation), in a stochastic or random process there is some indeterminacy: even if the initial condition (or starting point) is known, there are several (often infinitely many) directions in which the process may evolve.

In the simple case of discrete time, as opposed to continuous time, a stochastic process involves a sequence of random variables and the time series associated with these random variables (for example, see Markov chain, also known as discrete-time Markov chain).

Random field edit

Given a probability space   and a measurable space X, an X-valued random field is a collection of X-valued random variables indexed by elements in a topological space T. That is, a random field F is a collection

 

where each   is an X-valued random variable.

Several kinds of random fields exist, among them the Markov random field (MRF), Gibbs random field (GRF), conditional random field (CRF), and Gaussian random field. An MRF exhibits the Markovian property

 

where   is a set of neighbours of the random variable Xi. In other words, the probability that a random variable assumes a value depends on the other random variables only through the ones that are its immediate neighbours. The probability of a random variable in an MRF is given by

 

where Ω' is the same realization of Ω, except for random variable Xi. It is difficult to calculate with this equation, without recourse to the relation between MRFs and GRFs proposed by Julian Besag in 1974.

Random measure edit

A random measure is a measure-valued random element.[5][6] Let X be a complete separable metric space and   the σ-algebra of its Borel sets. A Borel measure μ on X is boundedly finite if μ(A) < ∞ for every bounded Borel set A. Let   be the space of all boundedly finite measures on  . Let (Ω, ℱ, P) be a probability space, then a random measure maps from this probability space to the measurable space ( ,  ).[7] A measure generally might be decomposed as:

 

Here   is a diffuse measure without atoms, while   is a purely atomic measure.

Random set edit

A random set is a set-valued random element.

One specific example is a random compact set. Let   be a complete separable metric space. Let   denote the set of all compact subsets of  . The Hausdorff metric   on   is defined by

 

  is also а complete separable metric space. The corresponding open subsets generate a σ-algebra on  , the Borel sigma algebra   of  .

A random compact set is а measurable function   from а probability space   into  .

Put another way, a random compact set is a measurable function   such that   is almost surely compact and

 

is a measurable function for every  .

Random geometric objects edit

These include random points, random figures,[8] and random shapes.[8]

References edit

  1. ^ Fréchet, M. (1948). "Les éléments aléatoires de nature quelconque dans un espace distancié". Annales de l'Institut Henri Poincaré. 10 (4): 215–310.
  2. ^ V.V. Buldygin, A.B. Kharazishvili. Geometric Aspects of Probability Theory and Mathematical Statistics. – Kluwer Academic Publishers, Dordrecht. – 2000
  3. ^ Yates, Daniel S.; Moore, David S; Starnes, Daren S. (2003). The Practice of Statistics (2nd ed.). New York: Freeman. ISBN 978-0-7167-4773-4. Archived from the original on 2005-02-09.
  4. ^ L. Castañeda; V. Arunachalam & S. Dharmaraja (2012). Introduction to Probability and Stochastic Processes with Applications. Wiley. p. 67. ISBN 9781118344941.
  5. ^ Kallenberg, O., Random Measures, 4th edition. Academic Press, New York, London; Akademie-Verlag, Berlin (1986). ISBN 0-12-394960-2 MR854102. An authoritative but rather difficult reference.
  6. ^ Jan Grandell, Point processes and random measures, Advances in Applied Probability 9 (1977) 502-526. MR0478331 JSTOR A nice and clear introduction.
  7. ^ Daley, D. J.; Vere-Jones, D. (2003). An Introduction to the Theory of Point Processes. Probability and its Applications. doi:10.1007/b97277. ISBN 0-387-95541-0.
  8. ^ a b Stoyan, D., and Stoyan, H. (1994) Fractals, Random Shapes and Point Fields. Methods of Geometrical Statistics. Chichester, New York: John Wiley & Sons. ISBN 0-471-93757-6

Literature edit

  • Hoffman-Jorgensen J., Pisier G. (1976) "Ann.Probab.", v.4, 587–589.
  • Mourier E. (1955) Elements aleatoires dans un espace de Banach (These). Paris.
  • Prokhorov Yu.V. (1999) Random element. Probability and Mathematical statistics. Encyclopedia. Moscow: "Great Russian Encyclopedia", P.623.

External links edit

  • Entry in Springer Encyclopedia of Mathematics