
Probabilistic Distributions
2026-04-01
A distribution is discrete if:
Examples of Discrete Outcomes
A discrete random variable is a numerical description of a countable outcome.
Typical forms:
Examples:
Discrete distributions are described by a Probability Mass Function (PMF).
The PMF answers the question:
What is the probability that the random variable equals a specific value?
Mathematically:
\[ P(X = x) \]
Core Properties of a PMF
\[ \sum_x P(X = x) = 1 \]
- Հավանականության խտության ֆունկից ա;
- Հավանականության զանգվածի ֆունկցիա
Poisson models turn randomness into operational clarity.
How many times does an event occur within a fixed time (or space) interval?
Imagine a retail store observing customer foot traffic.
Every 10 minutes, a random number of customers enter the store.
Over many days, management notices:
Yet the average stays roughly the same.
\[ P(X = x) = \frac{\lambda^x e^{-\lambda}}{x!}, \quad x = 0,1,2,\dots \]
This formula gives the probability of seeing exactly \(x\) events
\[ E[X] = \lambda \]
\[ Var(X) = \lambda \]
Poisson models are widely used in practice for:
They describe the same process from different perspectives.

The Bernoulli Distribution models the most fundamental probabilistic question in data analytics and business decision-making:
Did an event happen or not?
A Bernoulli random variable has only two possible outcomes:
These outcomes can be represented in multiple equivalent ways depending on context:
A random variable \(X\) follows a Bernoulli distribution if:
\[ X \sim \text{Bernoulli}(p) \]
where:
A Bernoulli trial is a single experiment with:
Each trial is assumed to be independent.
Common real-world Bernoulli trials include:
Each produces a binary outcome.
Consider an online store tracking customer conversions.
For each user session:
10Across thousands of sessions:
Each session is evaluated independently, with the same underlying probability of conversion.
Typical Bernoulli use cases in analytics:
Bernoulli is the building block for many advanced models.
Because Bernoulli is a discrete distribution, it is described by a Probability Mass Function (PMF):
\[ P(X = x) = p^x (1-p)^{1-x}, \quad x \in \{0,1\} \]
This formula yields exactly two probabilities:
For a Bernoulli random variable:
\[ E[X] = p \]
\[ Var(X) = p(1-p) \]
An e-commerce company records whether each visitor completes a purchase.
Each session is encoded as:
1 → purchase0 → no purchase{.smaller}
| Session | Purchase |
|---|---|
| 1 | 0 |
| 2 | 1 |
| 3 | 0 |
| 4 | 0 |
| 5 | 1 |
| 6 | 0 |
| 7 | 1 |
| 8 | 0 |
| 9 | 0 |
| 10 | 1 |
Let \(X\) be the purchase indicator per session.
Sample mean:
\[ \bar{x} = \frac{4}{10} = 0.4 \]
Estimated probability:
\[ \hat{p} = 0.4 \]
\[ X = \begin{cases} 1 & \text{if a purchase occurs} \\ 0 & \text{otherwise} \end{cases} \]
A Bernoulli model is an appropriate first representation of this process.
“Even if you’re not normal, the average is normal”
The Normal distribution appears everywhere in data.
The Central Limit Theorem states:
.
The CLT allows us to:
Even when the original data are not Normal.
Think about:
Individually, these can be far from Normal.
When we repeatedly take many samples and compute means:

Before applying CLT:
These ensure the sample mean behaves approximately Normal.
These methods assume approximate Normality of averages.
“Even if you’re not normal, the average is normal.”