Sie sind auf Seite 1von 2

Discrete probability distribution

Definition
In probability theory, a probability distribution is called discrete if it is characterized by a probability mass function. Thus, the distribution of a random variable X is discrete, and X is then called a discrete random variable, if

as u runs through the set of all possible values of X. If a random variable is discrete, then the set of all values that it can assume with nonzero probability is finite or countably infinite, because the sum of uncountably many positive real numbers (which is the least upper bound of the set of all finite partial sums) always diverges to infinity. Typically, this set of possible values is a topologically discrete set in the sense that all its points are isolated points. But, there are discrete random variables for which this countable set is dense on the real line. Among the most well-known discrete probability distributions that are used for statistical modeling are the Poisson distribution, the Bernoulli distribution, the binomial distribution, the geometric distribution, and the negative binomial distribution. In addition, the discrete uniform distribution is commonly used in computer programs that make equal-probability random selections between a number of choices.

Alternative description
Equivalently to the above, a discrete random variable can be defined as a random variable whose cumulative distribution function (cdf) increases only by jump discontinuities that is, its cdf increases only where it "jumps" to a higher value, and is constant between those jumps. The points where jumps occur are precisely the values which the random variable may take. The number of such jumps may be finite or countably infinite. The set of locations of such jumps need not be topologically discrete; for example, the cdf might jump at each rational number. Consequently, a discrete probability distribution is often represented as a generalized probability density function involving Dirac delta functions, which substantially unifies the treatment of continuous and discrete distributions. This is especially useful when dealing with probability distributions involving both a continuous and a discrete part.

Representation in terms of indicator functions


For a discrete random variable X, let u0, u1, ... be the values it can take with non-zero probability. Denote

These are disjoint sets, and by formula (1)

It follows that the probability that X takes any value except for u0, u1, ... is zero, and thus one can write X as

except on a set of probability zero, where 1A is the indicator function of A. This may serve as an alternative definition of discrete random variables.

Das könnte Ihnen auch gefallen