Probability Theory – Why Can a Probability Measure Be Defined Over Power Set of Countable Sample Space?

measure-theoryprobability theory

Let $\Omega$ be the sample space, $p: \Omega \rightarrow [0,1]$, be any function satisfying $\sum_{w\in\Omega} p(w) = 1$. Then there is a valid probability triple $(\Omega, \mathcal{F},P)$, where $\mathcal{F}$ is the collection of all subsets of $\Omega$ (the power set), and for $A \in \mathcal{F}, P(A) = \sum_{w\in A} p(w)$.

My question is why/how we know that, when $\Omega$ is finite or countable, a probability measure can be defined over ALL possible subsets (vs, say, the fact that a probability measure — satisfying the standard axioms, of course — cannot be defined over all subsets of [0,1]). Is there a proof of this somewhere (if so, please let me know where I can look), or could someone provide one?

Best Answer

You can always define certain measures on the whole power set, for example the dirac measure $\delta_x (A)=1$ if $x\in A $ and $\delta_x (A)=0$ otherwise.

Also, countable sums of Diracs work well. On a countable sample space, any measure is of this form (countable linear combination if Diracs).

Only when we consider more interesting measures like the Lebesgue measure, we can't define it on the whole powerset while maintaining desired properties like translation invariance.