5.2 Discrete Probability Distribution Function

Probability Distributions of Discrete Random Variables

Engineers often encounter systems and phenomena that exhibit inherent variability, whether it’s in manufacturing processes, environmental conditions, or communication channels. Understanding random variables is crucial for engineers as it provides them with the tools to model and analyze this variability mathematically. By studying random variables, engineers can develop probabilistic models that capture the uncertainties present in their systems, allowing for more robust designs, accurate predictions, and informed decision-making. Whether optimizing production processes, designing communication systems resilient to noise, or assessing structural safety in varying environmental conditions, a deep understanding of random variables equips engineers with the necessary framework to tackle real-world challenges effectively.

Random Variable Definition

A random variable is a rule or a function which assigns a numeric value to each of the possible outcomes of an experiment. If the outcomes are countable then the random variable is discrete. If the outcomes are measured then the random variable is continuous.

When we make a list of the possible values the random variable can take on along with the probability each value can occur, we create a probability distribution function, or PDF for short.

A discrete probability distribution function has two characteristics:

  1. Each probability is between zero and one, inclusive.
  2. The sum of the probabilities for all possible outcomes is one.

Discrete Probability Distribution Function

If X is a discrete random variable, and if for each possible outcome, x, of the random variable,

[latex]P(X = x) \geq 0[/latex]

[latex]\displaystyle{\sum_x} P(x)=1[/latex]

then the set of ordered pairs (x, P(x)) makes up a discrete probability distribution function, also called just the probability distribution or PDF.

Example 1 – Verify Discrete Probability Distribution Function

A child psychologist studying newborn babies, within the first seven days of life, and is interested in the number of times per week a newborn cries after midnight. For a random sample of 50 newborns, the following information was obtained. Verify the information in the table is a discrete probability distribution function.

Let the random variable X = the number of times per week a newborn baby cries after midnight. In this study, the babies observed cried between 0 and 5 times after midnight during a week.  Note: P(x) = probability that the random variable X takes on a value x.

x P(x)
0 P(x = 0) = 2/50
1 P(x = 1) = 11/50
2 P(x = 2) = 23/50
3 P(x = 3) = 9/50
4 P(x = 4) = 4/50
5 P(x = 5) = 1/50

X takes on the values 0, 1, 2, 3, 4, and 5. These are all positive counted values. The relative frequency or probability each occurred is listed as P(x). Note that if we add up each of the probabilities, we get 2/50 + 11/50 + 23/50 + 9/50 + 4/50 + 1/50 = 1.

This is a discrete PDF because:

  1. Each P(x) is between zero and one, inclusive.
  2. The sum of the probabilities is one.

 

A PDF is handy for a discrete random variable because we can calculate probabilities easily. In Example 1, the probability a randomly selected newborn will cry twice after midnight is 23/50 or 46%. What if we are curious about the probability a randomly selected baby cries at most twice? Note this means the baby can cry zero times, once, or twice, so all of the probability must be accumulated. We can accumulate probabilities to facilitate such calculations. We can add a column so that we accumulate all the probability at or below the value of the random variable.
x P(x) [latex]P(X \leq x)[/latex]
0 P(x = 0) = 2/50 [latex]P(X \leq 0)[/latex] = 2/50
1 P(x = 1) = 11/50 [latex]P(X \leq 1)[/latex] = 13/50
2 P(x = 2) = 23/50 [latex]P(X \leq 2)[/latex] = 36/50
3 P(x = 3) = 9/50 [latex]P(X \leq 3)[/latex] = 45/50
4 P(x = 4) = 4/50 [latex]P(X \leq 4)[/latex] = 49/50
5 P(x = 5) = 1/50 [latex]P(X \leq 0)[/latex] = 50/50 = 1
Now that we have a column of the cumulative probabilities, we can see that the probability a randomly selected baby cries at most twice is 36/50 or 72%. Note that by the time we exhausted the possible values of the random variable, in this case find the probability a randomly selected newborn cries at most five times after midnight, we accumulate 100% of the probability. It is important to avoid writing the fractions as decimals and then rounding the probabilities. This leads to rounding error.
The discrete probability distribution PDF can be represented by a probability histogram in which the bar heights correspond to the probabilities the random variable takes on that particular value. For the study in Example 1, the probability distribution histogram and the cumulative distribution histogram are provided.
In general, for any discrete random variable, the cumulative function can be found by adding up all the probabilities for all the possible values of the random variable that are less than or equal to a particular value.

Cumulative Distribution Function

For a discrete random variable X, with probability distribution given by P(X = x), the cumulative distribution function is given by

F(x) = [latex]\displaystyle{\sum_x}P(X \leq x)[/latex]

for each possible outcome x.

Inference in statistics involves drawing a sample from a population and studying the sample data so that conclusions and inference can be made about the population. In many situations, the probability density function can be well approximated or categorized as a standard known function. We will encounter several of these standard density functions in this class. Consider a situation in which there is one trial or one observation and there are exactly two outcomes. This is a very special case of a known random variable.

A Common Discrete Random Variable

A Bernoulli Random Variable or Bernoulli “trial” is a two option, “yes or no” random variable. The two options are often referred to as “success” or “failure” in the definition with one of them labeled as 1 and the other labeled as 0. It has one parameter p, which denotes the probability of observing the success. For engineers, the choice of assigning probabilities to success or failure is arbitrary. Often, however, in engineering a “success” is similar to the idea of “true”, which is labeled as 1 in any digital logic class. A Bernoulli random variable is named for Jacob Bernoulli who, in the late 1600s, studied them extensively.

Bernoulli Distribution

If X is a Bernoulli random variable, its probability distribution is

[latex]P(X = x)=\begin{cases}p & x = 1\\1 - p & x = 0\end{cases}[/latex]

Thus a Bernoulli random variable can take on only the values X = 0 or X = 1. The parameter p is the probability that the random variable is 1. This makes the probability that the random variable is 0 equal to 1 – p, so most often, we will see P(X = 1) = p and P(X = 0) = 1 – p. The random variable is said to have a Bernoulli Distribution with parameter, p. An often used notation for random variables involves naming them and placing the parameter in parentheses.

[latex]X \sim Bernoulli(p)[/latex]

When we flip one coin exactly one time, we are performing a Bernoulli trial. We can assign the observation of a heads as 1 or 0, depending on which one we want to call a success. If the coin is fair, and if we agree that observing tails is defined as a success, then [latex]X \sim Bernoulli(0.5)[/latex].

License

Icon for the Creative Commons Attribution 4.0 International License

Introduction to Statistics for Engineers Copyright © by Vikki Maurer & Jeff Crabill & Linn-Benton Community College is licensed under a Creative Commons Attribution 4.0 International License, except where otherwise noted.