Those four statements are indeed quite different! Towards unpacking their differences:
a) Notice that the processes in statements 1 and 3 are (essentially) the same-- they constitute at each point $\omega/t$ the partial sum of the observations that has jumps at the points of the form $i/n$. These are both basically then considering the partial sum process as an element of $D[0,1]$, the space of cadlag functions on $[0,1]$. Statement 1 is stronger than statement 3-- while statement 3 is basically saying that the distribution of the partial sum process is close to that of a Brownian motion, statement 1 is saying that there exists a copy of the original partial sum process, defined on a potentially new probability space, and Brownian motions defined on the same space, that are close in probability. As such, and it is a worthwhile exercise to consider, statement 1 can be used to prove statement 3 relatively easily, but not the other way around. Statement 1 belongs to a family of approximation results for stochastic processes known as "weak approximations", have a look at the Skorokhod-Dudley-Wichura theorem, and see https://encyclopediaofmath.org/wiki/Skorokhod_theorem. Note that while it seems kinda weird that all random variables must potentially be redefined on a new probability space, the necessity of doing so is for a very simple and understandable reason: the original sample space for the observations may simply not be rich enough to support a Brownian motion. Skorokhod's original proof works by constructing all variables on the sample space $(0,1)$ equipped with Lebesgue measure.
b) Statement two considers a modified partial sum process that, rather than having jumps, is continuously interpolated using linear interpolation. The processes in statement 1/3 and 2 agree on the points of the form $i/n$. The point of considering this process rather than the one in statement 1/3 is basically for mathematical convenience-- it takes values in the space $C[0,1]$ of continuous functions, which is a complete and separable metric space when equipped with the sup-norm $\|x-y\|=\sup_{t\in [0,1]}|x(t)-y(t)|$. Separability is a key tool in establishing many asymptotic results for measures defined on metric spaces. The space $D[0,1]$ equipped with the sup-norm is NOT separable. As developed in Chapter 3 of Billingsly's 1968 book, a metric on $D[0,1]$ can be defined, this is called the Skorokhod metric, making $D[0,1]$ separable, and such that many functionals of statistical/probabilisitc interest on $D[0,1]$ are continuous with respect to that metric, thereby circumventing the need to transform the partial sum process into $C[0,1]$, which admittedly is kind of clunky.
An even more slick way of handling this has been developed more recently, which is sometimes called weak convergence in Hoffman-Jorgensen sense. Basically in this case weak convergence is defined using outer expectation, and so processes that are not continuous, such as the standard partial sum process, can have their weak convergence considered in the metric space $C[0,1]$, since the weak limit, a Brownian motion, lives in this space. This theory is comprehensively developed in Vaart, Aad van der; Wellner, Jon A. Weak convergence and empirical processes.
c) Statement 4 is a statement about weak convergence of the standard empirical process, which is analogous to statement 3 for the partial sum process. Donsker's original papers on the topic consider these two cases separately, and the development of results in this vein since then have often followed this pattern.
Best Answer
There are various ways in which the CLT can "fail", depending on which hypotheses are violated. Here's one. Suppose $X_k$ are independent random variables with $E[X_k] = \mu_k$ and variances $\sigma_k^2$, and let $s_n^2 = \sum_{k=1}^n \sigma_k^2$ and $S_n = \sum_{k=1}^n (X_k - \mu_k)$. Suppose also that $\max_{k \le n} \sigma_k/s_n \to 0$ as $n \to \infty$ (so in that sense no $X_k$ is "dominant" in $S_n$). Then Lindeberg's condition is both necessary and sufficient for $S_n/s_n$ to converge in distribution to ${\mathscr N}(0,1)$.
EDIT: Here's a nice example where the Central Limit Theorem fails. Let $X_n$ be independent with $P(X_n = 2^n) = P(X_n = -2^n) = 2^{-2n-1}$, $P(X_n = 0) = 1 - 2^{-2n}$. Thus $E[X_n] = 0$ and $\sigma_n = 1$. But $$P(S_n = 0) \ge P(X_j = 0 \text{ for all }j) > 1 - \sum_{j=1}^\infty 2^{-2j} = 2/3$$