$$P\left(\bigcup_{m\ge1}\lim\sup_{k\to
> \infty}\{|X_k-X|\ge\frac{1}{m}\}\right)=0$$
Which is equivalent to $\underline{\exists\epsilon>0} $ such that
$$P\left(\lim\sup_{k\to \infty}\{|X_k-X|\ge\epsilon\}\right)=0 $$
(Note I fixed a misplaced brace in the first statement, and changed $n$ to $k$ which I think is what you mean.)
Those are not equivalent.
What is true is that the former would be equivalent to
$$P\left(\exists \epsilon > 0 : \limsup_{n\to
> \infty}\{|X_k-X|\}\ge\epsilon\right)=0.$$
This says, more precisely, that the following set has measure zero: the set of all $\omega \in \Omega$ for which there exists $\epsilon > 0$ such that $$\limsup_{n\to
> \infty}\{|X_k(\omega)-X(\omega)|\}\ge \epsilon.$$
Note the order of quantifiers: $\epsilon$ is allowed to depend on $\omega$.
In your second statement, you are asking that there be a single $\epsilon$ that works for almost every $\omega$ simultaneously. That is a much stronger condition, and so it is much easier for the corresponding set of $\omega$ to be null.
To see the problem more vividly, consider the sequence $X_k = 1$ for all $k$, and $X=0$. Using your first statement, we can see that we do not have $X_k \to X$ a.s, because for every $m > 1$ and every $k$, the set $\{|X_k - X| \ge \frac{1}{m} \}$ is all of $\Omega$. So the $\limsup_{k \to \infty}$ of these sets is also $\Omega$ for all $m > 1$, and so the union over all $m$ is also $\Omega$. Thus the probability is $1$.
But using the second statement, suppose I take $\epsilon = 2$. Then the set $\{|X_k - X| \ge \epsilon\}$ is empty for all $k$, so the limsup is also empty. The probability of the empty set is 0. So your second statement is satisfied in this example, even though $X_k$ does not converge to $X$ a.s.
Update
In your current numbering, (1), (2), (3), (4), (6) are all equivalent to one another, and (5) is weaker.
To see why (1) is equivalent to (2), for notational convenience let's write $$A_\epsilon = \limsup_{k \to \infty} \{|X_k - X| <\epsilon\} = \bigcup_{n \ge 1} \bigcap_{k \ge n} \{|X_k - X| < \epsilon\}.$$ Then (1) reads $\forall \epsilon > 0 \, P(A_\epsilon) = 1$, and (2) reads $P(\bigcap_{m \ge 1} A_{1/m}) = 1$.
Suppose that (1) holds. Then $P(A_{1/m}) = 1$ for all $m \ge 1$. By countable additivity, it follows that $P(\bigcap_{m \ge 1} A_{1/m}) = 1$ as well, so we have (2). (One way to see this is that $(\bigcap_{m \ge 1} A_{1/m})^c = \bigcup_{m \ge 1} A_{1/m}^c$ which is a countable union of measure-zero sets.)
Conversely, suppose that (2) holds. Given $\epsilon > 0$, choose $M$ so large that $1/M < \epsilon$, and note that $A_{1/M} \subseteq A_\epsilon$. On the other hand, $\bigcap_{m \ge 1} A_{1/m} \subset A_{1/M}$ since $A_{1/M}$ is one of the sets being intersected on the left. So by monotonicity of measure, we have
$$1 = P\left(\bigcap_{m \ge 1} A_{1/m}\right) \le P(A_{1/M}) \le P(A_\epsilon)$$
and thus $P(A_\epsilon) = 1$. Since $\epsilon > 0$ was arbitrary, (1) is proved.
To look at (4) and (5), set
$$B_\epsilon = A_\epsilon^c = \limsup_{k \to \infty} \{|X_k - X| \ge \epsilon\} = \bigcap_{n \ge 1} \bigcup_{k \ge n} \{|X_k - X| \ge \epsilon\}.$$
Then your (4) reads $P(\bigcup_{m \ge 1} B_{1/m}) = 0$ and (5) is $\exists \epsilon > 0 \, P(B_\epsilon) = 0$.
We do have (4) implies (5): take $\epsilon = 1$. Then $B_1 \subset \bigcup_{m \ge 1} B_{1/m}$, so $P(B_1) \le P(\bigcup_{m \ge 1} B_{1/m}) = 0$. But (5) does not imply (4); see the counterexample I gave above.
Why can't you prove (5) implies (4) in a similar way to the proof that (1) implies (2)? Just try and you'll see why it doesn't work. Maybe you know that one of the sets $B_{1/m}$ has measure zero (if you are lucky and the $\epsilon$ that works in (5) is smaller than 1). But we have $B_1 \subseteq B_{1/2} \subseteq B_{1/3} \subseteq \dots$ and so knowing that one of them has measure zero does not tell you anything about the measure of the later sets in the sequence.
You might like to compare the following statements. Suppose $C_1, C_2, \dots$ is a sequence of events.
(a) If $P(C_n) = 1$ for every $n$, then $P(\bigcap_n C_n) = 1$. (TRUE)
(b) If $P(C_n) = 0$ for every $n$, then $P(\bigcup_n C_n) = 0$. (TRUE)
(c) If $P(C_n) = 0$ for some $n$, then $P(\bigcup_n C_n) = 0$. (FALSE)
You were probably thinking of something like (c) when you thought that (5) would imply (4). You might think that "symmetry" should suggest (a) implies (c), but it clearly doesn't.
Best Answer
I don't really see intuition here, the equivalence just follows from using the definition of convergence. For a sequence of sets $(A_n)$ the set $\lim \sup(A_n)=\{A_n\ \ i.o\}$ is the set of elements which belong to infinitely many of the sets $A_n$. The formal definition of this set is $\cap_{n=1}^\infty \cup_{k=n}^\infty A_k$.
Assume $X_n\to X$ almost surely by the first definition and let any constant $\epsilon>0$. Define the sequence $A_{n,\epsilon}:=\{\omega: |X_n(\omega)-X(\omega)|>\epsilon\}$. Note that if $\omega\in\lim\sup A_{n,\epsilon}$ then it means that $|X_n(\omega)-X(\omega)|>\epsilon$ for infinitely many values of $n$, and hence $X_n(\omega)$ obviously does not converge to $X(\omega)$. So $\lim\sup A_{n,\epsilon}\subseteq \{\omega: X_n(\omega)\nrightarrow X(\omega)\}$, and by monotonicity of probability:
$\mathbb{P}(\lim\sup A_{n,\epsilon})\leq \mathbb{P}(\{\omega: X_n(\omega)\nrightarrow X(\omega)\})=0$
Second direction: Now assume $X_n\to X$ by the second definition. For each $k\in\mathbb{N}$ define $B_k=\lim\sup A_{n,\frac{1}{k}}$ where the sets $A_{n,\epsilon}$ are defined like before. Then by assumption $\mathbb{P}(B_k)=0$ for all $k$, and hence $\mathbb{P}(\cup_{k=1}^\infty B_k)=0$. Now suppose we have $X_n(\omega)\nrightarrow X(\omega)$ for some $\omega$. This implies that there must be some $m\in\mathbb{N}$ such that $|X_n(\omega)-X(\omega)|>\frac{1}{m}$ for infinitely many natural numbers $n$, and thus $\omega\in B_m\subseteq\cup_{k=1}^\infty B_k$.
In other words, we have the inclusion $\{\omega: X_n(\omega)\nrightarrow X(\omega)\}\subseteq\cup_{k=1}^\infty B_k$, and so $\mathbb{P}(\{\omega: X_n(\omega)\nrightarrow X(\omega)\})=0$.