Summarize this article:
253 LearnersLast updated on November 20, 2025

An important concept in statistics and data science, the Bernoulli distribution is one of the ways to describe an event with only two possible outcomes. The outcomes are represented as ‘success’ or ‘failure’ or ‘1’ and ‘0.’ Let us understand the concept with an example. Suppose we participated in a race, we either win or lose. The Bernoulli distribution helps explain situations like these. In this topic, we will explore the concept of Bernoulli distribution.
The Bernoulli Distribution is the mathematical framework for the simplest kind of experiment: one where there are only two possible ways things can end.
Think of it as the “Either/Or” model. In statistical language, we typically label these two outcomes as 1 (Success) and 0 (Failure).
This concept, known as a Bernoulli trial, appears frequently in daily life. You are running a Bernoulli trial every time you:
The Bernoulli distribution is famous not because it's complex, but because it is the foundation for much larger statistical concepts. It is the “atom” of discrete probability.
Without Bernoulli, these more complex models wouldn't exist.
Example:
Imagine you are rolling a standard die. Let's say you decide that landing on Six counts as a “success.”
While the math is simple, this model gives statisticians a standardized way to describe binary outcomes in everything from gambling to manufacturing.
Learning the properties of Bernoulli distribution helps students understand the situations with only two possible outcomes. It is a simple way to know how probability works in real life and also in subjects, such as economics, mathematics, statistics, and computer science. The properties of Bernoulli distribution are listed below:
The Bernoulli distribution formula expresses the probability of two possible results. It shows the chances of success and failure.
The Bernoulli distribution has two important functions, probability distribution function (PDF) and cumulative distribution function (CDF). The PDF gives the probability of success and failure. The CDF, on the other hand, describes the probability of getting a value less than or equal to 1 or 0.
The PDF for Bernoulli distribution is:
\(P (X = x) = px (1 − p)1 − x\)
Here,
\(P (X = x) =\) The chance of getting the possible outcome, like success or failure.
\(X =\) The outcome
Where \(x = 1\) means success and \(x = 0\) means failure
\(p =\) Probability of success.
\(1 - p =\) Probability of failure.
Now, let us take a closer look at this formula with an example:
In a running competition, if the probability of winning is 0.7, then the probability of losing is:
\(1 − 0.7 = 0.3\)
By using the Bernoulli distribution formula, we can find the probability.
Probability of success:
\( P (X = x) = px (1 − p) − x \)
\( P (X = 1) = (0.7) (0.3) − 1 \)
\(0.7 × 1 = 0.7\)
So, the probability of winning is 0.7
Probability of failure:
\( P (X = x) = px (1 − p) − x \)
\( P (X = 0) = (0.7) (0.3)1 − 0 \)
\(1 × 0.3 = 0.3\)
Therefore, the probability of losing is 0.3
The cumulative distribution function for the Bernoulli distribution is:
\(F(x) = 0\), when \(x < 0\).
\(F(x) = 1 − p\), if \(0 ≤ x < 1\), and
\(F(x) = 1\), if \(x ≥ 1\).
\(F(x) =\) The cumulative distribution function and the random variable \(X\) is less than or equal to \(x\).
\(P (X ≤ x) =\) Probability of \(X\) is less than or equal to \(x\).
Let us see what are the differences between a Bernoulli Distribution and a Binomial Distribution.
|
Feature |
Bernoulli Distribution |
Binomial Distribution |
|
Definition |
A probability distribution for a single trial with two possible outcomes (Success or Failure). |
A probability distribution for the number of successes in a sequence of n independent Bernoulli trials. |
|
Number of Trials (n) |
Always \(n = 1\). |
A fixed number, n (where \(n \ge 1\)). |
|
Possible Outcomes (X) |
Only two: 0 (Failure) or 1 (Success). |
A range of integers: \(\{0, 1, 2, \dots, n\}.\) |
|
Parameters |
\(p\)(probability of success). |
\(n\)(number of trials) and \(p\)(probability of success). |
|
Notation |
\(X \sim \text{Bernoulli}(p)\) |
\(X \sim \text{Binomial}(n, p)\) |
|
PMF Formula (Probability Mass Function) |
\(P(X=x) = p^x (1-p)^{1-x}\), (Where\( x \in \{0,1\}\)) |
\(P(X=k) = \binom{n}{k} p^k (1-p)^{n-k}\), (Where \(k\) is the number of successes) |
|
Mean (Expected Value) |
\(\mu = p\) |
\(\mu = np\) |
|
Variance |
\(\sigma^2 = p(1-p)\) |
\(\sigma^2 = np(1-p)\) |
|
Real-World Example |
Flipping a coin once. (Heads or Tails) |
Flipping a coin 10 times and counting how many were Heads. |
The mean represents the expected average outcome for a distribution, telling you what value to expect over the long run. Variance measures the distribution of uncertainty around the mean, quantifying how likely the actual results are to deviate from the expected average.
The Probability Mass Function (PMF) of a Bernoulli distribution specifies the probability that a random variable X takes on a specific value x. Since the Bernoulli distribution only has two possible outcomes (Success or Failure), the PMF is quite simple. It is typically expressed as:
\(P(X=x) = \begin{cases} p & \text{if } x = 1 \text{ (Success)} \\ 1-p & \text{if } x = 0 \text{ (Failure)} \\ 0 & \text{otherwise} \end{cases}\)
The Cumulative Distribution Function (CDF) calculates the probability that a random variable X will take a value less than or equal to a specific number x. For a Bernoulli distribution (where outcomes are strictly 0 or 1), the CDF is a step function that increases in jumps at \(x=0\) and \(x=1\).
\(F(x) = P(X \le x) = \begin{cases} 0 & \text{if } x < 0 \\ 1 - p & \text{if } 0 \le x < 1 \\ 1 & \text{if } x \ge 1 \end{cases}\)
Where:
The Bernoulli distribution graphs depict a single binary experiment, with the PMF showing two distinct bars representing the probabilities of failure and success. This is paired with the CDF, a staircase-like step function that depicts the cumulative probability, rising at each outcome until it reaches 100 percent.
Probability Mass Function (PMF): This graph shows the probabilities of the two possible outcomes. You will see two distinct bars.
Cumulative Distribution Function (CDF): This graph shows the running total of probability. It looks like a staircase.
A Bernoulli trial is an experiment with only two possible outcomes: success or failure. It is named after the mathematician James Bernoulli.
Common Examples:
The 3 Essential Rules:
The Bernoulli distribution is the simplest probability model with only two outcomes, success or failure. Mastering it builds the foundation for understanding more complex probability distributions.
Students learning about the properties, concepts, and real-life applications of the Bernoulli distribution can apply it to benefit their academic pursuits as well as to tackle real-world scenarios. Sometimes, misunderstanding the concepts of the Bernoulli distribution can lead to wrong decisions and incorrect calculations. By understanding some common errors and their solutions will help students improve their problem-solving skills and analytical capabilities.
The concept of Bernoulli distribution can be applied to predict the outcome of events. Probability plays a major role even in our day-to-day lives. Here are some examples where it is used.
A factory produces light bulbs, and 3% of them are defective. If a bulb is picked randomly, what is the probability that it is defective or non-defective?
3% chance of picking a defective and a 97% chance of picking a non-defective bulb.
Here, the formula is:
\(P (X = x) = px (1 − p)1 − x\)
(X = 1) = If the bulb is defective
(X = 0) = If the bulb is non-defective
Probability of success \((p) = 3\%\) or \(0.03 (3 / 100 = 0.03)\)
The defective bulbs \((X = 1)\):
\(P (X = 1) = (0.03)1 × (1 − 0.03)0\)
\(P (X = 1) = 0.03 × 1 = 0.03\)
For non-defective bulbs \((X = 0)\):
\(P (X = 0) = (0.03)0 × (1 − 0.03)1\)
Since (0.03)0 = 1
\(P (X = 0) = 1 × 0.97 = 0.97\)
Hence, the probability of picking a defective bulb is 0.03, and the probability of picking a non-defective bulb is 0.97.
A student attends school 80% of the time on any given day. What is the probability that the student attends or misses school on a randomly selected day?
80% chance of attending school and 20% chance of missing school
Here,
Success \((X = 1)\): The student attends school
Failure \((X = 0)\): The student misses school
Probability of success: 80% or 0.8
Now let us apply the formula:
\(P (X = x) = px (1 − p)1 − x\)
\( P (X = 1) = (0.8)1 × (1 − 0.8) \)
\(P (X = 1) = 0.8 × 1 = 0.8\)
For missing school (X = 0):
\( P (X = 0) = (0.8) × (1 − 0.8) = 0.2 \)
The student has an 80% (0.8) chance of attending school and a 20% (0.2) chance of missing school.
A washing machine has a 70% chance of working when switched on. What is the probability that it fails?
0.3 (30%)
The probability of success \((p) = 0.7 (70\%)\)
Probability of failure \((1 − p) = 0.3\)
To find the probability of failure:
\(1 × 0.3 = 0.3\)
Therefore, the probability of failure is 0.3 (30%).
A student has a 50% chance of passing an exam. What is the probability that they fail?
50%
Success = passing the exam
Probability of success \((p) = 0.5\)
Probability of failure \((1 − p) = 0.5\)
To find the probability of failure (X = 0):
\( P (X = 0) = (0.5)^0 × (0.5)^1 \)
\(P (X = 0) = 1 × 0.5 = 0.5\)
So, the probability of failing the exam is 0.5 or 50%.
A weather forecast says there is a 60% chance of rain today. What is the probability it won’t rain?
40% of probability.
The probability of rain \((p) = 0.6\)
The probability of no rain \((1 − p) = 0.4\)
The formula for the probability of no rain is:
\( P (X = 0) = (0.6)^0 × (0.4)^1 \)
\(P (X = 0) = 1 × 0.4 = 0.4\)
Hence, the probability of no rain is 40% or 0.4.
Jaipreet Kour Wazir is a data wizard with over 5 years of expertise in simplifying complex data concepts. From crunching numbers to crafting insightful visualizations, she turns raw data into compelling stories. Her journey from analytics to education ref
: She compares datasets to puzzle games—the more you play with them, the clearer the picture becomes!






