Maybe it helps if we first widen our view, in order to then narrow it again and see the double-dual as special case.
So let's start with functions (any functions, for now) $f:X\to Y$. Let's as concrete example, take $X=Y=\mathbb R$. That is, we are dealing with real-values functions of a real argument. Examples would be the identity $\mathrm{id} = x\mapsto x$, the constant functions $\mathrm{const}_c = x\mapsto c$, of the trigonometric functions $\sin$ and $\cos$.
Now the normal way to look at functions is to think of them as encoding the operation, for example, it is a property of the function $\sin$ that it maps the number $\pi$ to the number $0$:
$$\sin(\pi) = 0$$
But another view is that the result of applying the function $\sin$ to the number $\pi$ gives the number $0$, and it is that applying that has all the logic. So you have one function $\mathrm{apply}$ that takes two arguments, a real function and a real number, and assigns them another number:
$$\mathrm{apply}(\sin,\pi)=0$$
Now looking at this form, we see that $\sin$ and $\pi$ are on equal footing. Both are merely arguments of the $\mathrm{apply}$ function. You recover the original sine function by “pre-inserting” $\sin$ as first argument of apply (this is known as currying):
$$x\mapsto \mathrm{apply}(\sin,x)$$
But given that both arguments are on equal footing, you may just as well pre-apply the second argument instead:
$$f\mapsto \mathrm{apply}(f,\pi)$$
We might consider this the application of $\pi$ to the function $f$. Thus $\mathrm{apply}(\sin,\pi)$ could equivalently be written as
$$\pi(\sin) = 0$$
So now from each real number, we get a function that maps real functions to real numbers. Note that just like the function $\sin$ is not determined just by the value $\sin(\pi)$, but by the values it takes for all real numbers, similarly, the function $\pi$ is not determined just by the value it takes at $\sin$, but by the values it takes for all real functions. That is, we not only have $\pi(\sin)=0$, but also $\pi(\cos)=-1$, $\pi(\mathrm{id})=\pi$ and $\pi(\mathrm{const_c})=c$.
Note also that the real functions form an $\mathbb R$-vector space under pointwise addition and scalar multiplication. And it is easily determined that those “number functions” defined above are linear functions, that is, they live in the dual space of that function space. And quite obviously they only form a proper subset of that dual space, as they for example don't include the constant function $f\mapsto 0$ (as there is no real number that is mapped to $0$ by all real functions). Indeed, that example shows that here we don't even have a subspace here.
However we have an injection into that dual, as we can identify each number by looking only at the function values. Easiest of course by applying it to the identity function (that returns the number itself), but even if we did not have that available (as will be the case below), we could e.g. look at the functions that are $1$ for exactly one number, and $0$ for all others; with those functions we can uniquely identify the number by just noting whioch of those functions give a value of $1$.
Now let's look instead at a vector space $V$ over a field $K$, and at linear functions $V\to K$, that is, members of the dual $V^*$. Again, we can do the same game as above, and for each vector, we get a function mapping members of $V^*$ to the dual of $V^*$, which is the double dual of $V$.
However, now that we have only linear functions, we get more than above: The function that maps vectors to members of the double dual can easily be shown to be linear itself. And again, we can construct a set of functions in $V^*$ that uniquely identifies the vector: Choose a basis $\{b_i\}$ in $V$, and then take the set of linear functions $f_i$ that map $v = \sum_i\alpha_i b_i$ to $\alpha_i$. Since a vector is uniquely identified by its basis coefficients, this proves that the map $V\to V^{**}$ is injective: You can uniquely identify the vector by the values $v(f_i)=\alpha_i$.
Best Answer
I think about quotient spaces in the following way.
Let $V$ be a vector space and $W$ be a subspace. Then the vector space $V/W$ is simply the vector space $V$ in which you have killed every element of $W$. Thus, $V/W$ can be considered as the vector space $V$ with the extra rule that an element of $W$ becomes $0$.
Let's see how this intuition applies in the situation of the isomorphism theorem.
Let $T: V \to V'$ be a linear map. The isomorphism theorem tells us that
$$V/ \ker T \cong \operatorname{Im} T.$$
Why do we intuitively expect this? Well, $V/\ker T$ is the vector space $V$ in which we kill the kernel, i.e. an element of the kernel becomes $0$ in our vector space. So basically this means that our map becomes injective! Thus, we get an isomorphism on the image (injection on the image of a map is bijective).