One of the axioms of a vector space is that $\lambda \vec{v} \in V$ for all $\lambda \in \mathbb{F}$ and $\vec{v} \in V$.
Let $V=\mathbb{R}$ and $\mathbb{F}=\mathbb{C}$. If $\vec{v}=1$ and $\lambda=\operatorname{i}$ then $\lambda \vec{v} = \operatorname{i}$ and that does not belong to $\mathbb{R}$.
In such a case, the pair $(\mathbb{R},\mathbb{F})$ fails to be closed under scalar multiplication.
First, not all functions from the reals to the reals are "polynomial, triginometric, exponential". These form in fact a very tiny subset of all functions. A function is not a formula. It's just that for every $x \in \mathbb{R}$ there exists a unique value $f(x) \in \mathbb{R}$. The unicity is what makes it a function. The asignment of $f(x)$ to $x$ is completely arbitrary, e.g. I could take the 7-th digit in the decimal (infinite) representation of $x$ for all $x > \sqrt{17}$, the 5-th digit of that representation for all $x < \sqrt{17}$ and $f(\sqrt{17}) = \pi$. And this is even relatively nice, because I can write a program for it, but this need not be the case in general. A truly "random" function would just be an infinite list of values, one for each real, assigned all independently of each other. So there is no hope for Taylor expansions etc. Realise that $\mathbb{R}^\mathbb{R}$ is truly a huge set.
The operations on $\mathbb{R}^\mathbb{R}$ are pointwise, so if we have two such functions $f,g$ then we define $f+g$ as a new function, by telling what its value on an arbitrary $x \in \mathbb{R}$ is: $(f+g)(x) = f(x) + g(x)$, i.e. we just add the values of $f$ and $g$ at $x$. Similarly for a scalar $c \in \mathbb{R}$, we define $(c\cdot f)(x) = cf(x)$ for all $x \in \mathbb{R}$, where the latter is just standard multiplication in $\mathbb{R}$. The $0$ is just the function where all values are equal to $0$.
Both have infinite sets of linearly independent elements (or vectors, as elements of a vector space are called, even though they are not "vectors" in the old fashioned sense, like the functions in $\mathbb{R}^\mathbb{R}$): take the functions $f_p$, defined for a fixed $p \in \mathbb{R}$: $f_p(x) = 1$ if $x = p$, and $f_p(x) = 0$ if $x \neq p$. So all $0$ except for a spike at $p$.
Why are the $f_p$ linearly independent? By definition, we need to consider a finite linear combination of distinct $f_{p_i}$: $c_{p_1} \cdot f_{p_1} + \ldots + c_{p_n} \cdot f_{p_n} = 0$ (equality as functions, which means just that they have the same values for all $x$), and we need to show that then all $c_{p_i}$ are $0 \in \mathbb{R}$. Because the equality holds for all $x$, we can use $x = p_1$ in particular. Then $$(c_{p_1} \cdot f_{p_1} + \ldots + c_{p_n} \cdot f_{p_n})(p_1) = c_{p_1}f_{p_1}(p_1) + \ldots + c_{p_n} f_{p_n}(p_1) = 0$$ As $f_{p_2}(p_1) = 0$, as $p_2 \neq p_1$, and so on, but $f_{p_1}(p_1) = 1$ this comes down to $c_{p_1} = 0$. The same idea works for all other coefficients as well. So the set of $f_p$ is linearly independent.
The same idea works in $\mathbb{R}^\mathbb{N}$, the set of sequences, which look more like normal vectors, but of infinite length. We just see this as the set of functions from $\mathbb{N}$ to $\mathbb{R}$ and the same operations and independent functions apply. So both spaces are infinite-dimensional.
If we see $\mathbb{R}^\mathbb{N}$ as a set of functions (as we should) then the function $T$ is just the restriction of $f$ to $\mathbb{N}$. To see that this is linear, take $f,g \in \mathbb{R}^\mathbb{R}$. Then $T(f+g)$ is defined for all $n$ as $(f+g)(n)$, which is in turn defined as $f(n) + g(n)$, and this equals $T(f)(n) + T(g)(n)$, as $T$ does "nothing", it's just the restriction of a function to a smaller domain. The latter sum is just by definition $(T(f) + T(g))(n)$, and as this holds for all $n$ we have equality of functions (or sequences, because we index by $\mathbb{N}$) and $T(f+g) = T(f) + T(g)$. The same can be done for scalars as well.
A matrix for a linear map $T$ is formed by choosing bases for both spaces and computing the base expansion for every $T(b)$ for all basis elements in the domain (as the columns). But here a base for $\mathbb{R}^\mathbb{R}$ is uncountable, so we cannot write it down as a matrix: these can be at most countably infinite in both dimensions.
Best Answer
There is nothing wrong with defining a vector space over a field $F$ whose elements are $n$-tuples with entries from a field $K$, where $K$ contains $F$.
A simple check will show that all the axioms of vector space are satisfied.
To answer what is the meaning of $\mathbb C^n$ being a vector space over $\mathbb R$, consider the special case of $n=1$. One can visualize the complex numbers as arrows tailed at origin.
For $n>1$ this cannot be visualized and so the question of 'meaning' cannot be answered.
Mathematical objects can be very abstract and 'meaningless'.
As to how to differentiate between the two vector spaces in the two problems: One is a vector space and the other isn't!