Bernoulli Distribution: Mean & Variance in Business Stats
Explore the Bernoulli distribution in business statistics. Understand its mean, variance, and key concepts for AI and machine learning applications.
Bernoulli Distribution in Business Statistics: Mean and Variance
This document provides a comprehensive overview of the Bernoulli distribution, focusing on its fundamental concepts, formulas, and applications in business statistics, with a particular emphasis on its mean and variance.
11.1 Terminologies Associated with Bernoulli Distribution
The Bernoulli distribution is a discrete probability distribution that describes the outcome of a single Bernoulli trial. Key terminologies include:
- Bernoulli Trial: A random experiment with exactly two possible outcomes, typically labeled as "success" and "failure."
- Success: The outcome of interest in a Bernoulli trial.
- Failure: The outcome that is not of interest in a Bernoulli trial.
- Probability of Success ($p$): The probability of achieving the "success" outcome in a Bernoulli trial.
- Probability of Failure ($q$): The probability of achieving the "failure" outcome in a Bernoulli trial. By definition, $q = 1 - p$.
11.2 Formula of Bernoulli Distribution
The probability mass function (PMF) of a Bernoulli distribution is defined as follows:
$P(X=x) = \begin{cases} p & \text{if } x = 1 \text{ (success)} \ 1-p & \text{if } x = 0 \text{ (failure)} \end{cases}$
Where:
- $X$ is the random variable representing the outcome of a Bernoulli trial.
- $x$ is the value of the random variable (either 0 for failure or 1 for success).
- $p$ is the probability of success.
This can be more compactly written as:
$P(X=x) = p^x (1-p)^{1-x}$, for $x \in {0, 1}$
11.3 Mean and Variance of Bernoulli Distribution
The mean (expected value) and variance are key measures of a Bernoulli distribution's central tendency and spread, respectively.
11.3.1 Mean (Expected Value)
The mean of a Bernoulli distribution represents the average outcome over many trials. It is calculated as:
$E(X) = \sum_{x} x P(X=x)$ $E(X) = (1 \times p) + (0 \times (1-p))$ $E(X) = p$
Therefore, the mean of a Bernoulli distribution is equal to the probability of success ($p$).
11.3.2 Variance
The variance measures the dispersion of the distribution around its mean. It is calculated as:
$Var(X) = E[(X - E(X))^2]$ $Var(X) = \sum_{x} (x - E(X))^2 P(X=x)$ $Var(X) = (1 - p)^2 p + (0 - p)^2 (1-p)$ $Var(X) = (1 - 2p + p^2)p + p^2 (1-p)$ $Var(X) = p - 2p^2 + p^3 + p^2 - p^3$ $Var(X) = p - p^2$ $Var(X) = p(1-p)$
Alternatively, using the formula $Var(X) = E(X^2) - [E(X)]^2$:
$E(X^2) = (1^2 \times p) + (0^2 \times (1-p)) = p$ $Var(X) = p - p^2 = p(1-p)$
Therefore, the variance of a Bernoulli distribution is $p(1-p)$.
11.4 Properties of Bernoulli Distribution
- Discrete: The random variable can only take on a finite number of values (0 and 1).
- Two Outcomes: There are only two possible results for the trial.
- Independent Trials: Each Bernoulli trial is independent of previous trials.
- Constant Probability of Success: The probability of success ($p$) remains the same for every trial.
- Relationship to Binomial Distribution: A Bernoulli distribution is a special case of the binomial distribution where the number of trials ($n$) is 1.
11.5 Bernoulli Distribution Graph
The graph of a Bernoulli distribution typically consists of two bars, representing the probabilities of success (1) and failure (0).
- X-axis: Represents the outcome ($x=0$ for failure, $x=1$ for success).
- Y-axis: Represents the probability ($P(X=x)$).
The height of the bar at $x=1$ will be $p$, and the height of the bar at $x=0$ will be $1-p$.
^ Probability
|
p +-------+
| |
| |
(1-p) +-------+
| |
+-------+-------+------> Outcome
0 1
11.6 Bernoulli Trial
A Bernoulli trial is the fundamental concept underlying the Bernoulli distribution. It is a single experiment or event that meets the following criteria:
- Exactly two possible outcomes: These outcomes are mutually exclusive and exhaustive.
- Probabilities are fixed: The probability of each outcome remains constant across all trials.
- Independence: The outcome of one trial does not affect the outcome of any other trial.
Examples:
- Flipping a fair coin once (Heads/Tails).
- A single customer purchasing a product (Yes/No).
- A manufactured item passing a quality inspection (Pass/Fail).
11.7 Examples of Bernoulli Distribution
-
Coin Flip: Consider flipping a fair coin. Let "Heads" be a success (X=1) and "Tails" be a failure (X=0). The probability of success $p = 0.5$.
- $P(X=1) = 0.5$
- $P(X=0) = 1 - 0.5 = 0.5$
- Mean: $E(X) = p = 0.5$
- Variance: $Var(X) = p(1-p) = 0.5(0.5) = 0.25$
-
Customer Conversion: A marketing team runs an ad campaign. For a single user, the probability that they click the ad is $p = 0.1$. Let "click" be a success (X=1) and "no click" be a failure (X=0).
- $P(X=1) = 0.1$
- $P(X=0) = 1 - 0.1 = 0.9$
- Mean: $E(X) = p = 0.1$
- Variance: $Var(X) = p(1-p) = 0.1(0.9) = 0.09$
11.8 Applications of Bernoulli Distribution in Business Statistics
The Bernoulli distribution serves as a foundational building block for analyzing many business scenarios involving binary outcomes:
- Marketing and Sales:
- Probability of a customer clicking an advertisement.
- Probability of a sales lead converting into a sale.
- Probability of a customer making a repeat purchase.
- Quality Control:
- Probability of a product passing inspection.
- Probability of a component being defective.
- Risk Management:
- Probability of a loan defaulting.
- Probability of an insurance claim being filed.
- Customer Service:
- Probability of a customer being satisfied with a service.
- Probability of a customer abandoning a support call.
- Operations:
- Probability of a machine breaking down in a given period.
- Probability of a specific event occurring in a process.
11.9 Difference between Bernoulli Distribution and Binomial Distribution
While related, the Bernoulli and Binomial distributions differ in the number of trials considered:
Feature | Bernoulli Distribution | Binomial Distribution |
---|---|---|
Number of Trials ($n$) | Exactly 1 trial | $n$ independent trials (where $n \ge 1$) |
Outcomes | Two outcomes: Success (1) or Failure (0) | Exactly two outcomes for each trial: Success or Failure |
Random Variable ($X$) | $X=1$ if success, $X=0$ if failure. | $X$ = the total number of successes in $n$ trials. |
Parameters | $p$ (probability of success) | $n$ (number of trials) and $p$ (probability of success) |
PMF | $P(X=x) = p^x (1-p)^{1-x}$, for $x \in {0, 1}$ | $P(X=k) = \binom{n}{k} p^k (1-p)^{n-k}$, for $k = 0, 1, ..., n$ |
Mean | $E(X) = p$ | $E(X) = np$ |
Variance | $Var(X) = p(1-p)$ | $Var(X) = np(1-p)$ |
Relationship | A Binomial distribution with $n=1$ is a Bernoulli distribution. | A sum of $n$ independent Bernoulli trials follows a Binomial distribution. |
In essence, the Bernoulli distribution describes the outcome of a single event, while the Binomial distribution describes the number of successes in a series of identical and independent Bernoulli trials.
Bivariate Frequency Distribution Limitations in AI
Explore the key limitations of bivariate frequency distributions in AI & ML. Understand constraints for effective data analysis and model selection.
Bernoulli Distribution: Key Terminologies for ML & AI
Master Bernoulli distribution with essential terminologies like success, failure, binary outcomes, and trials. Crucial for understanding AI/ML probability concepts.