Definitions for Continuous Random Variables
In the previous chapter we mainly focused on discrete random variables whose set of possible values is either finite or countably infinite. In this chapter, we study random variables whose set of possible values is uncountable. We’ll see later on that a lot of the cases we’ve discussed have analogs in the continuous case.
Probability density function
The continuous random variable is a random variable with infinite possible outcomes (a subset of the real line). We say $X$ is a continuous random variable
if there exists a nonnegative function $f$ defined for all $x \in (-\infty, \infty)$, having the property that for any set $B$ of real numbers,
$$
P\{x \in B\} = \int_B f(x)dx
$$
Here the function $f$ is called the probability density function
(PDF). It resembles the probability mass function in the discrete case. The PDF has the following properties:
- $\int_{-\infty}^\infty f(x)dx = P\{X \in (-\infty, \infty)\} = 1$
- $P\{a \leq x \leq b\} = \int_a^b f(x)dx$
- $P\{X = a\} = \int_a^a f(x)dx = 0$
We can also define the cumulative distribution function
for a continuous random variable:
$$
\begin{aligned}
F(a) &= P\{X \leq a\}, \quad B \in (-\infty, a] \\
&= \int_{-\infty}^a f(x)dx
\end{aligned}
$$
Properties example
These properties often come in handy when we have unknown quantities in a PDF. Suppose $X$ is a continuous random variable with probability density function $$ f(x) = \begin{cases} C\left( 4x - 2x^2 \right), & 0 < x < 2 \\ 0, & \text{otherwise} \end{cases} $$ and we’d like to find $C$ as well as the probability $P\{X > 1\}$. $$ \begin{aligned} 1 &= \int_{-\infty}^\infty f(x)dx \\ &= \int_{-\infty}^0 f(x)dx + \int_0^2 f(x)dx + \int_2^\infty f(x)dx \\ &= C\int_0^2 (4x - 2x^2)dx \\ &= C\left(\int_0^2 4xdx - \int_0^2 2x^2dx \right) \\ &= C \left( 2x^2\bigg|_0^2 - \frac{2}{3}x^3 \bigg|_0^2 \right) \\ &= C\left( 8 - \frac{16}{3} \right) \end{aligned} $$ Now we can easily find $C = \frac{3}{8}$. With the PDF given, it’s trivial to find the CDF:
$$ \begin{aligned} P{X > 1} &= \int_1^\infty f(x)dx \\ &= \int_1^2 \frac{3}{8}\left(4x - 2x^2\right)dx \\ &= \frac{3}{8}\left( 2x^2\bigg|_1^2 - \frac{2}{3}x^3 \bigg|_1^2 \right) \\ &= \frac{3}{8}\left( 8-2 - \left(\frac{16}{3} - \frac{2}{3}\right) \right) = \frac{1}{2} \end{aligned} $$
Lifetime example
Suppose $X$, the lifetime of an item, is a continuous random variable with a density function $$ f(x) = \begin{cases} \lambda e^{-x/100}, & x \geq 0 \\ 0, & x < 0 \end{cases} $$ What is the probability that the item functions between 50 and 150 days?
In mathematical terms, we want to calculate $P\{50 \leq X \leq 150\}$. We first need to find the value of $\lambda$.
$$ \begin{aligned} 1 &= \int_{\infty}^\infty f(x)dx \\ &= \int_0^\infty \lambda e^{-\frac{x}{100}}dx \\ &= \lambda \int_0^\infty e^{-\frac{x}{100}}dx \\ &= -100\lambda \int_0^\infty e^{-\frac{x}{100}} d\frac{-x}{100} \\ &= -100\lambda e^{-\frac{x}{100}} \bigg|_0^\infty \\ &= -100\lambda(0 - 1) = 100\lambda \end{aligned} $$
Recall that $\int e^xdx = e^x$, and $d(ax) = a \cdot dx$ because the derivative is a linear function.
With $\lambda = \frac{1}{100}$, we can calculate
$$ \begin{aligned} P{50 \leq X \leq 150} &= \int_{50}^{150} \frac{1}{100}e^{-\frac{x}{100}}dx \\ &= \frac{-100}{100}\int_{50}^{150}e^{-\frac{x}{100}} d\frac{-x}{100} \\ &= -\left( e^{-\frac{x}{100}}\bigg|_{50}^{150} \right) \\ &= e^{-\frac{1}{2}} - e^{-\frac{3}{2}} \approx 0.383 \end{aligned} $$
Expectation and variance
Earlier we’ve defined the expectation for discrete random variables. If $X$ is a continuous random variable with probability density function $f(x)$, we have $$ f(x)dx \approx P\{x \leq X \leq x + dx\} $$
so it’s easy to find the analog for the expectation
of $X$ to be
$$ E[X] = \int_{-\infty}^\infty xf(x)dx $$
Similarly, the expected value of a real-valued function of $X$ is
$$ E[g(x)] = \int_{-\infty}^\infty g(x)f(x)dx $$
which can be used to derive the variance
of $X$
$$
Var(X) = E[X^2] - E[X]^2
$$
Suppose $X$ is a continuous random variable with density function $$ f(x) = \begin{cases} 2x, & 0 \leq x \leq 1 \\ 0, & \text{otherwise} \end{cases} $$
$$ \begin{aligned} E[X] &= \int_{-\infty}^\infty xf(x)dx \\ &= \int_0^1 x \cdot 2x dx \\ &= \frac{2}{3}x^3 \bigg|0^1 = \frac{2}{3} \\ E[X^2] &= \int{-\infty}^\infty x^2 f(x)dx \\ &= \int_0^1 2x^3 dx \\ &= \frac{1}{2}x^4 \bigg|_0^1 = \frac{1}{2} \\ Var(X) &= \frac{1}{2} - \left(\frac{2}{3}\right)^2 = \frac{1}{18} \end{aligned} $$
Next, we introduce some commonly seen continuous probability distributions.
Dec 28 | Sampling Distribution and Limit Theorems | 6 min read |
Dec 08 | Functions of Random Variables | 14 min read |
Nov 06 | Multivariate Probability Distributions | 23 min read |
Nov 01 | Common Continuous Random Variables | 17 min read |
Oct 06 | Definitions for Discrete Random Variables | 7 min read |