[Math] Linear transformation between infinite dimensional vector space

linear algebralinear-transformationsmatricesvector-spaces

So the question asks:

Let $\Bbb R^{\Bbb R}$ be the vector space of all functions $f : \Bbb R\to \Bbb R$, and let $\Bbb R^{\Bbb N}$ be the vector space of all infinite sequences, or tuples, $(x_1, x_2,\ldots , x_n,\ldots)$ of real numbers.

(a) Show that both $\Bbb R^{\Bbb R}$ and $\Bbb R^{\Bbb N}$ are infinite-dimensional.

(b) Prove that the function $T : \Bbb R^{\Bbb R}\to\Bbb R^{\Bbb N}$ given by $T(f) = (f(1), f(2),\ldots)$ is a linear transformation.Why doesn’t it make sense to talk about the matrix of the linear transformation T?

So so far I have:

For (a) ,

Since the functions from $\Bbb R$ to $\Bbb R$ are polynomial, trigonometric, exponential and etc., by Taylor series expansion, every function has a polynomial representation of either finite degree or infinite degree.

So, the set of all functions over real numbers can be written as

$\Bbb R^{\Bbb R} = \{a_0+ a_1x + a_2x_2+\ldots+ a_nx_n +\ldots\mid a_i\in\Bbb R\}$

So, each function can be written as n-tuples where n is finite or infinite as

$a_0+ a_1x + a_2x_2+\ldots+ a_nx_nn+\ldots=(a_1, a_2,\ldots, a_n, \ldots)$

So, $\Bbb R^{\Bbb R}=\{(a_1, a_2,\ldots, a_n,\ldots)\mid a_i\in\Bbb R)\}$

This forms a vector space under addition of $n$– tuples and scalar multiplication.

Therefore, the function from $\Bbb R$ to $\Bbb R$ forms an infinite dimensional vector space $\Bbb R^{\Bbb R}$.

The space of all infinite sequences of real numbers can be written as

$\Bbb R^{\Bbb N}=\{(a_1, a_2,\ldots, a_n,\ldots)\mid a_i\in\Bbb R\}$.

Each sequence can be written as a linear combination of

$S = \{(1,0,0,\ldots,0,\ldots),(0,1,0,\ldots,0,\ldots),(0,0,1,0,\ldots,0,\ldots),\ldots\}$

So there are infinite vectors in $S$ which span $\Bbb R^{\Bbb N}$

Therefore, $\Bbb R^{\Bbb N}$ is an infinite dimensional vector space.

For (b),

I know to prove a transformation is linear, it has to satisfy:

  1. $T(v_1+v_2)=T(v_1)+T(v_2)\forall v_1, v_2\in V$, and

  2. $T(kv)=kT(v)$ for any scalar $k$.

So, I need to prove that

  1. $T(f_1) +T(f_2) = T(f_1+f_2)$

  2. $T(k\cdot f_1) = k T(f1)$

But, how am I supposed to do this? And the question asks why it makes no sense to talk about the matrix of $T$, is it because the transformation is from an infinite-dimensional vector space to the another infinite-dimensional vector space, and the matrix is abstract which has infinite basis?

Best Answer

First, not all functions from the reals to the reals are "polynomial, triginometric, exponential". These form in fact a very tiny subset of all functions. A function is not a formula. It's just that for every $x \in \mathbb{R}$ there exists a unique value $f(x) \in \mathbb{R}$. The unicity is what makes it a function. The asignment of $f(x)$ to $x$ is completely arbitrary, e.g. I could take the 7-th digit in the decimal (infinite) representation of $x$ for all $x > \sqrt{17}$, the 5-th digit of that representation for all $x < \sqrt{17}$ and $f(\sqrt{17}) = \pi$. And this is even relatively nice, because I can write a program for it, but this need not be the case in general. A truly "random" function would just be an infinite list of values, one for each real, assigned all independently of each other. So there is no hope for Taylor expansions etc. Realise that $\mathbb{R}^\mathbb{R}$ is truly a huge set.

The operations on $\mathbb{R}^\mathbb{R}$ are pointwise, so if we have two such functions $f,g$ then we define $f+g$ as a new function, by telling what its value on an arbitrary $x \in \mathbb{R}$ is: $(f+g)(x) = f(x) + g(x)$, i.e. we just add the values of $f$ and $g$ at $x$. Similarly for a scalar $c \in \mathbb{R}$, we define $(c\cdot f)(x) = cf(x)$ for all $x \in \mathbb{R}$, where the latter is just standard multiplication in $\mathbb{R}$. The $0$ is just the function where all values are equal to $0$.

Both have infinite sets of linearly independent elements (or vectors, as elements of a vector space are called, even though they are not "vectors" in the old fashioned sense, like the functions in $\mathbb{R}^\mathbb{R}$): take the functions $f_p$, defined for a fixed $p \in \mathbb{R}$: $f_p(x) = 1$ if $x = p$, and $f_p(x) = 0$ if $x \neq p$. So all $0$ except for a spike at $p$.

Why are the $f_p$ linearly independent? By definition, we need to consider a finite linear combination of distinct $f_{p_i}$: $c_{p_1} \cdot f_{p_1} + \ldots + c_{p_n} \cdot f_{p_n} = 0$ (equality as functions, which means just that they have the same values for all $x$), and we need to show that then all $c_{p_i}$ are $0 \in \mathbb{R}$. Because the equality holds for all $x$, we can use $x = p_1$ in particular. Then $$(c_{p_1} \cdot f_{p_1} + \ldots + c_{p_n} \cdot f_{p_n})(p_1) = c_{p_1}f_{p_1}(p_1) + \ldots + c_{p_n} f_{p_n}(p_1) = 0$$ As $f_{p_2}(p_1) = 0$, as $p_2 \neq p_1$, and so on, but $f_{p_1}(p_1) = 1$ this comes down to $c_{p_1} = 0$. The same idea works for all other coefficients as well. So the set of $f_p$ is linearly independent.

The same idea works in $\mathbb{R}^\mathbb{N}$, the set of sequences, which look more like normal vectors, but of infinite length. We just see this as the set of functions from $\mathbb{N}$ to $\mathbb{R}$ and the same operations and independent functions apply. So both spaces are infinite-dimensional.

If we see $\mathbb{R}^\mathbb{N}$ as a set of functions (as we should) then the function $T$ is just the restriction of $f$ to $\mathbb{N}$. To see that this is linear, take $f,g \in \mathbb{R}^\mathbb{R}$. Then $T(f+g)$ is defined for all $n$ as $(f+g)(n)$, which is in turn defined as $f(n) + g(n)$, and this equals $T(f)(n) + T(g)(n)$, as $T$ does "nothing", it's just the restriction of a function to a smaller domain. The latter sum is just by definition $(T(f) + T(g))(n)$, and as this holds for all $n$ we have equality of functions (or sequences, because we index by $\mathbb{N}$) and $T(f+g) = T(f) + T(g)$. The same can be done for scalars as well.

A matrix for a linear map $T$ is formed by choosing bases for both spaces and computing the base expansion for every $T(b)$ for all basis elements in the domain (as the columns). But here a base for $\mathbb{R}^\mathbb{R}$ is uncountable, so we cannot write it down as a matrix: these can be at most countably infinite in both dimensions.