STATS 101

Right, What the Hell is a Probability Distribution Anyway?

KnowStatistics Jan 12, 2026 8 min read
A chart showing a discrete and a continuous probability distribution.
Figure: Mapping out all possible outcomes and their chances.

Introduction: Seriously, What Does This Even Mean?

Let's be honest. You've probably heard the term "probability distribution" bandied about. You hear it in news reports, you read it in articles, and you just nod along. But inside, you're thinking, "What the hell is that, really?" It's one of those terms that sounds incredibly clever but feels completely meaningless. You hear it, but you just don't get it.

If that sounds like you, this guide is everything you need from the A to the Z. Forget the white noise. At its core, a probability distribution is just a map. It's a simple rulebook, a table, or a chart that does two simple jobs:

  1. It lists all possible outcomes of a random event.
  2. It tells you how likely (the probability) each of those outcomes is.

That's it. It doesn't tell you what will happen next. It just lays out what could happen, and the chances for each. It's how we get a grip on uncertainty.

Before we dive in, we need one simple bit of terminology: a Random Variable. It sounds formal, but it's just a variable (which we usually call $X$) whose value is the numerical outcome of a random process. For example:

  • If we roll a die, we can let $X$ = the number that comes up.
  • If we flip a coin 3 times, we can let $X$ = the total number of heads.
  • If we measure a random student's height, we can let $X$ = their height in centimetres.

A probability distribution is simply the map that tells us the probability for every possible value of $X$. And it turns out, these maps come in two main flavours.

Flavour 1: Discrete Distributions (Things You Count)

The first flavour is the most intuitive. A discrete random variable is one that has a finite, or "countable," number of outcomes. You can list all the possibilities, even if the list is long. Think of things you count.

  • Number of heads in 2 coin flips: $X$ can be 0, 1, or 2. (Countable: 3 outcomes)
  • Number on a six-sided die: $X$ can be 1, 2, 3, 4, 5, or 6. (Countable: 6 outcomes)
  • Number of customers arriving in an hour: $X$ can be 0, 1, 2, 3... (Countable: infinite, but still in distinct steps)

The probability distribution for a discrete variable is often just a table or a bar chart. For a fair six-sided die, it's this simple:

Outcome ($x$) Probability $P(X=x)$
1 1/6 (or 16.7%)
2 1/6 (or 16.7%)
3 1/6 (or 16.7%)
4 1/6 (or 16.7%)
5 1/6 (or 16.7%)
6 1/6 (or 16.7%)

That's it! That table is the probability distribution. It lists all possible outcomes and their chances. Notice two iron-clad rules for all discrete distributions:

  1. The probability of any single outcome $x$ must be between 0 and 1 (i.e., $0 \le P(X=x) \le 1$).
  2. All the probabilities must add up to exactly 1. (Something must happen!)

The coin flip example is a bit more complex, but it follows the same rules. If you flip 3 fair coins, the distribution for $X$ = "Number of Heads" looks like this:

  • $P(X=0)$ (TTT) = 1/8
  • $P(X=1)$ (HTT, THT, TTH) = 3/8
  • $P(X=2)$ (HHT, HTH, THH) = 3/8
  • $P(X=3)$ (HHH) = 1/8

Again, all probabilities are between 0 and 1, and they all add up: $1/8 + 3/8 + 3/8 + 1/8 = 8/8 = 1$. This specific shape is called a Binomial Distribution, and it's one of the most famous discrete distributions.

Discrete Distribution Explorer: Coin Flips

See how the probability distribution (the theoretical long-run outcome) changes as you add more coins. Then, run a simulation to see how 1,000 "real" trials compare.

This bar chart shows the theoretical probability of getting $k$ heads. Notice the shape?

Flavour 2: Continuous Distributions (Things You Measure)

This is where things get a little more abstract, but just as intuitive. A continuous random variable is one that can take on any value within a range. There are infinite possibilities. Think of things you measure.

  • A person's exact height: $X$ could be 170.1cm, 170.12cm, 170.12345... cm.
  • The time you wait for a bus: $X$ could be 5.1 minutes, 5.123... minutes.
  • The temperature outside.

Here's the problem: because there are infinite possible outcomes, the probability of any single, exact outcome is zero. What's the probability that a person is exactly 170.123456789... cm tall? It's zero. It's impossible.

So, we change the question. For continuous variables, we never ask for the probability of a single value. We always ask for the probability of a range.

  • What's the probability a person is between 170cm and 171cm?
  • What's the probability you'll wait less than 5 minutes?
  • What's the probability the temperature is above 25°C?

We can't use a bar chart for this. Instead, we use a smooth curve called a Probability Density Function (PDF). This curve is drawn so that the area under the curve is the probability.

With discrete variables, $P(X=x)$ is the height of the bar.
With continuous variables, $P(a < X < b)$ is the area under the curve.

This PDF curve has its own set of rules, similar to the discrete ones:

  1. The curve can never go below the x-axis (no negative probabilities).
  2. The total area under the entire curve must be exactly 1. (Just like all the bars had to add to 1).

You've already seen the most famous continuous distribution: the Normal Distribution, or bell curve! But there are many other shapes, too.

Continuous Distribution Explorer

Continuous distributions are defined by a "Probability Density Function" (PDF) curve. The total area under each curve is 1 (100%). The probability of an outcome falling in a certain range is the area of that slice.

Select a distribution type to see its shape.

The "Centre" of Uncertainty: Expected Value

In descriptive statistics, we used the mean to find the "centre" of our data. How do we find the "centre" of a probability distribution? We use a concept called Expected Value, written as $E(X)$.

The expected value is a "weighted average" of all possible outcomes, where each outcome is weighted by its probability. It's the average result you would expect to get if you ran the random event an infinite number of times.

For a discrete variable, the formula is simple:

$E(X) = \sum [x \cdot P(x)]$

That just means: "For every outcome, multiply its value ($x$) by its probability ($P(x)$), then add all those products together."

Let's use our fair 6-sided die example:

  • $1 \cdot (1/6) = 1/6$
  • $2 \cdot (1/6) = 2/6$
  • $3 \cdot (1/6) = 3/6$
  • $4 \cdot (1/6) = 4/6$
  • $5 \cdot (1/6) = 5/6$
  • $6 \cdot (1/6) = 6/6$

Add them up: $E(X) = 1/6 + 2/6 + 3/6 + 4/6 + 5/6 + 6/6 = 21/6 = 3.5$.

Notice something funny? The expected value, 3.5, isn't even a possible outcome! You can't roll a 3.5. That's fine. It just means that if you rolled the die thousands of times, the average of all your rolls would get extremely close to 3.5. It's the "balance point" of the distribution, just like the mean. For continuous distributions, the $E(X)$ is also the balance point of the PDF curve (for a symmetrical curve like the Normal, $E(X)$ is right in the middle).

Guess the Expected Value

The expected value $E(X)$ is the 'balance point' of the distribution. A random discrete distribution will appear. Use the slider to guess its expected value.

Why This Is the Key to... Everything

This might all seem a bit theoretical, but probability distributions are the engine of data science, finance, insurance, and science. They allow us to model uncertainty in a mathematical way, which lets us make predictions and manage risk.

  • An insurance company uses distributions of "car accidents" or "house fires" to set premiums so they don't go bankrupt.
  • A business models the "distribution of daily sales" to know how much inventory to stock. Stock too much, and money is wasted. Stock too little, and you lose sales.
  • A scientist models the "distribution of measurement error" to know if their experiment's results are real or just random noise.
  • A finance professional models the "distribution of stock returns" to build a portfolio that balances risk and reward.

Without a probability distribution, you're just guessing. With one, you're making an informed decision based on a mathematical model of all possible futures.

Conclusion: From Chance to Inference

You've made another huge leap! You now understand that a probability distribution is just a map of uncertainty. It lists all possible outcomes and their probabilities, allowing us to model random events.

You learned that these distributions come in two main flavours: discrete (for things you count, like coin flips, represented by bar charts) and continuous (for things you measure, like height, represented by density curves where area = probability). You also learned how to find the "centre" or "balance point" of a distribution using its Expected Value.

This is the foundation for the next, most crucial step in statistics. So far, we've been talking about theoretical distributions, like a perfectly fair coin or a perfectly Normal curve. These are the "population" models. But in the real world, we almost never see the whole population. We just have a small sample.

The next big question is: How does a sample (like the mean of 30 students) relate to the population (all students)? What happens when we start taking samples over and over? This leads us to the single most important idea in all of statistics: Sampling Distributions and the Central Limit Theorem. This is the bridge that will finally allow us to use a small sample to make a confident inference about the entire world.

Probability Probability Distribution Random Variable Expected Value Statistics 101
Author Photo

About KnowStatistics

Hello, I'm Nina. I'm the founder, writer, and designer behind KnowStatistics. I spend a lot of my time pondering the whys, and chances are, you've stumbled upon an idea I once had.

I believe statistics to be the key to understanding the world, and it is imperative that it is accessible - so that's the heart of KnowStatistics.

Comments