In probability theory, **random element** is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by Maurice Fréchet (1948) who commented that the “development of probability theory and expansion of area of its applications have led to necessity to pass from schemes where (random) outcomes of experiments can be described by number or a finite set of numbers, to schemes where outcomes of experiments represent, for example, vectors, functions, processes, fields, series, transformations, and also sets or collections of sets.”^{[1]}

The modern-day usage of “random element” frequently assumes the space of values is a topological vector space, often a Banach or Hilbert space with a specified natural sigma algebra of subsets.^{[2]}

Let be a probability space, and a measurable space. A **random element** with values in *E* is a function *X*: Ω→*E* which is -measurable. That is, a function X such that for any , the preimage of B lies in .

Sometimes random elements with values in are called -valued random variables.

Note if , where are the real numbers, and is its Borel σ-algebra, then the definition of random element is the classical definition of random variable.

The definition of a random element with values in a Banach space is typically understood to utilize the smallest -algebra on *B* for which every bounded linear functional is measurable. An equivalent definition, in this case, to the above, is that a map , from a probability space, is a random element if is a random variable for every bounded linear functional *f*, or, equivalently, that is weakly measurable.

Main article: Random variable |

A *random variable* is the simplest type of random element. It is a map is a measurable function from the set of possible outcomes to .

As a real-valued function, often describes some numerical quantity of a given event. E.g. the number of heads after a certain number of coin flips; the heights of different people.

When the image (or range) of is finite or countably infinite, the random variable is called a discrete random variable^{[3]} and its distribution can be described by a probability mass function which assigns a probability to each value in the image of . If the image is uncountably infinite then is called a continuous random variable. In the special case that it is absolutely continuous, its distribution can be described by a probability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous,^{[4]} for example a mixture distribution. Such random variables cannot be described by a probability density or a probability mass function.

Main article: Random vector |

A **random vector** is a column vector (or its transpose, which is a row vector) whose components are scalar-valued random variables on the same probability space , where is the sample space, is the sigma-algebra (the collection of all events), and is the probability measure (a function returning each event's probability).

Random vectors are often used as the underlying implementation of various types of aggregate random variables, e.g. a random matrix, random tree, random sequence, random process, etc.

Main article: Random matrix theory |

A **random matrix** is a matrix-valued random element. Many important properties of physical systems can be represented mathematically as matrix problems. For example, the thermal conductivity of a lattice can be computed from the dynamical matrix of the particle-particle interactions within the lattice.

Main article: Random function |

A random function is a type of random element in which a single outcome is selected from some family of functions, where the family consists some class of all maps from the domain to the codomain. For example, the class may be restricted to all continuous functions or to all step functions. The values determined by a random function evaluated at different points from the same realization would not generally be statistically independent but, depending on the model, values determined at the same or different points from different realisations might well be treated as independent.

Main article: Random process |

A **Random process** is a collection of random variables, representing the evolution of some system of random values over time. This is the probabilistic counterpart to a deterministic process (or deterministic system). Instead of describing a process which can only evolve in one way (as in the case, for example, of solutions of an ordinary differential equation), in a stochastic or random process there is some indeterminacy: even if the initial condition (or starting point) is known, there are several (often infinitely many) directions in which the process may evolve.

In the simple case of discrete time, as opposed to continuous time, a stochastic process involves a sequence of random variables and the time series associated with these random variables (for example, see Markov chain, also known as discrete-time Markov chain).

Main article: Random field |

Given a probability space and a measurable space X,
an *X*-valued random field is a collection of *X*-valued
random variables indexed by elements in a topological space *T*. That is, a random field *F* is a collection

where each is an *X*-valued random variable.

Several kinds of random fields exist, among them the Markov random field (MRF), Gibbs random field (GRF), conditional random field (CRF), and Gaussian random field. An MRF exhibits the Markovian property

where is a set of neighbours of the random variable *X*_{i}. In other words, the probability that a random variable assumes a value depends on the other random variables only through the ones that are its immediate neighbours. The probability of a random variable in an MRF is given by

where Ω' is the same realization of Ω, except for random variable *X*_{i}. It is difficult to calculate with this equation, without recourse to the relation between MRFs and GRFs proposed by Julian Besag in 1974.

Main article: Random measure |

A **random measure** is a measure-valued random element.^{[5]}^{[6]} Let X be a complete separable metric space and the σ-algebra of its Borel sets. A Borel measure μ on X is boundedly finite if μ(A) < ∞ for every bounded Borel set A. Let be the space of all boundedly finite measures on . Let (Ω, ℱ, *P*) be a probability space, then a random measure maps from this probability space to the measurable space (, ).^{[7]} A measure generally might be decomposed as:

Here is a diffuse measure without atoms, while is a purely atomic measure.

A random set is a set-valued random element.

One specific example is a random compact set. Let be a complete separable metric space. Let denote the set of all compact subsets of . The Hausdorff metric on is defined by

is also а complete separable metric space. The corresponding open subsets generate a σ-algebra on , the Borel sigma algebra of .

A **random compact set** is а measurable function from а probability space into .

Put another way, a random compact set is a measurable function such that is almost surely compact and

is a measurable function for every .

These include random points, random figures,^{[8]} and random shapes.^{[8]}