I'm quite confused by the notion of random variable in the proper measure-theoretic framework. Let's first state the notation and definitions:
Let $(\Omega, \Sigma, \operatorname{P})$ be a probability space. Then, a real-valued random variable is a measurable function $X \colon \Omega \to \mathbb{R}$ and its probability distribution is the pushforward measure $\operatorname{P}_{X} := \operatorname{P} \circ X^{-1}$. If $\operatorname{P}_{X}$ is absolutely continuous with respect to the Lebesgue measure $\lambda$ we also know that there is a probability density function $f\colon \mathbb{R} \to \mathbb{R}$ such that $\operatorname{P}_{X}(B) = \int_B f \, \mathrm{d} \lambda$ for $B \in \mathcal{B}(\mathbb{R})$ (by the Radon–Nikodym theorem).
Now let's see a simple example that is often used to illustrate the notion of random variable:
- Random variable that represents the sum of two dice. In this case $\Omega = \{1, 2, 3, 4, 5, 6\}^2$, $\Sigma = \mathcal{P}(\Omega)$, and $\operatorname{P}(A) = \frac{\#A}{36}$ for $A \in \Sigma$, $X \colon (\omega_1, \omega_2) \mapsto \omega_1 + \omega_2$ and e.g. $\operatorname{P}_X(3) = \operatorname{P}(\{(1, 2), (2, 1)\}) = \frac{1}{18}$.
This is all crystal clear but the two examples below break my little mind:
- Normal random variable. What is $(\Omega, \Sigma, \operatorname{P})$ now? Others have given the answer that the underlying probability space is just abstract and unspecified. But why then, is it necessary to use the notion of random variable in the first place here? Wouldn't it be easier just to say that we are working with a probability space with $\Omega = \mathbb{R}$, $\Sigma = \mathcal{B}(\mathbb{R})$, and $\operatorname{P}(A) = \int_A \frac{1}{\sigma \sqrt{2 \pi}} e^{-\frac{1}{2}\left(\frac{x - \mu}{\sigma}\right)^2}\, \mathrm{d}\lambda(x)$ for $A \in \Sigma$?
- Random variable that represents the outcome of the toss of a fair coin. As explained here, the underlying probability space is again some abstract space of all conceivable futures. But why do we even need that? Why not directly use $\Omega = \{0, 1\}$, $\Sigma = \mathcal{P}(\Omega)$, and $\operatorname{P}(A) = \frac{\#A}{2}$ for $A \in \Sigma$?
If it is indeed beneficial to introduce random variables in these two cases, what are the benefits?