Probability Density Function (PDF)

For a huge class of probability distributions or probability measures, the probability of an event (i.e., the random variable falling into a certain set A) is found by integrating a specific function—the probability density function (PDF), denoted f(x)—over that set. The "integration" itself is done with respect to the Lebesgue measure (λ).

A probability measure P is absolutely continuous with respect to the Lebesgue measure λ if it can be represented as:

P(A)=Af(x)dλ(x)

This is usually written in the more familiar elementary calculus notation:

P(A)=Af(x)dx

For f(x) to be a valid PDF, it must satisfy two conditions:

  1. Non-negativity: f(x)0 for all x.
  2. Normalization: The total integral over the entire space must be 1.
Rf(x)dx=1

Think of the Lebesgue measure dx as providing the fundamental notion of "length" on the real line. The PDF f(x) then acts as a weighting factor, stretching or shrinking that length to assign a probability.

Famous Examples

1. Normal (Gaussian) Distribution

The famous "bell curve" is the PDF for the Normal distribution. To find the probability that a normally distributed random variable X falls between a and b, you integrate its PDF:

2. Exponential Distribution

Used to model waiting times, its probability is also the integral of its PDF.

3. Uniform Distribution on [a,b]

This is the simplest case. The PDF is a constant over the interval, meaning the probability is directly proportional to the length of the sub-interval (its Lebesgue measure).

It's crucial to know that not all distributions work this way.

Discrete Distributions

Discrete distributions, like the Poisson or Binomial, are singular with respect to the Lebesgue measure.

Their probability mass is concentrated on a countable set of points (e.g., the integers {0,1,2,...}). This set has a Lebesgue measure of zero. So, you cannot find a PDF f(x) to integrate.

Instead, they are defined by a probability mass function (PMF), which gives the probability at each specific point. In the language of measure theory, a discrete distribution is a weighted sum of Dirac measures.