[Math] Why is the eigenvalue equation $ Ax = \lambda x $ nonlinear

eigenvalues-eigenvectorslinear algebralinear-transformations

I read in a book that the eigenvalue equation is nonlinear but I can't see why that is form just looking at it…

$$
Ax = \lambda x
$$

looks a lot similar to

$$
Ax = b
$$

and if the matrix $A$ represents a linear transformation of $x$ then I don't see why $Ax = \lambda x$ would be nonlinear…

Best Answer

It sounds like you're interested in why "equation problems" $Ax=b$ are called linear while "eigenvalue problems" $Ax=\lambda\cdot x$ are called nonlinear. There are likely many reasons to use this language, but here's one that invokes the notion of a linear combinations.

Suppose that $x_1$ and $x_2$ are two solutions to "equation problems" $Ax_1=b_1$ and $Ax_2=b_2$. Consider an arbitrary linear combination $x$ of $x_1$ and $x_2$, so $x=c_1\cdot x_1+c_2\cdot x_2$. If we define $b=c_1\cdot b_1+c_2\cdot b_2$, then $$ Ax = A(c_1\cdot x_1+c_2\cdot x_2) = c_1\cdot Ax_1+c_2\cdot Ax_2 = c_1\cdot b_1+c_2\cdot b_2 = b $$ This illustrates that "linear combinations of solutions to equation problems are solutions to equation problems."

Now, suppose that $x_1$ and $x_2$ are two solutions to "eigenvalue problems" $Ax_1=\lambda_1\cdot x_1$ and $Ax_2=\lambda_2\cdot x_2$. Again consider an arbitrary linear combination $x=c_1\cdot x_1+c_2\cdot x_2$. This linear combination $x$ solves an eigenvalue problem if $Ax=\lambda\cdot x$ for some $\lambda$. However, we have $$ Ax = A(c_1\cdot x_1+c_2\cdot x_2) = c_1\cdot Ax_1+c_2\cdot Ax_2 = c_1\cdot\lambda_1\cdot x_1+c_2\cdot\lambda_2\cdot x_2 \overset{?}{=} \lambda\cdot(c_1\cdot x_1+c_2\cdot x_2)=\lambda\cdot x $$ The $?$ in this equation indicates that a suitable $\lambda$ might not exist. Indeed, it isn't difficult to find exmamples where such a $\lambda$ does not exist. For instance, consider the data \begin{align*} A &= \left[\begin{array}{rr} 23 & 32 \\ -16 & -25 \end{array}\right] & \lambda_1 &= 7 & x_1 &= \left[\begin{array}{r} 2 \\ -1 \end{array}\right] & \lambda_2 &= -9 & x_2 &= \left[\begin{array}{r} 1 \\ -1 \end{array}\right] \end{align*} Note that $Ax_1=\lambda_1\cdot x_1$ and $Ax_2=\lambda_2\cdot x_2$. However, for $x=x_1+x_2$, we have $$ \overset{A}{\left[\begin{array}{rr} 23 & 32 \\ -16 & -25 \end{array}\right]}\overset{x}{\left[\begin{array}{r} 3 \\ -2 \end{array}\right]} = \left[\begin{array}{r} 5 \\ 2 \end{array}\right] \neq \lambda\cdot x $$

Side Note. As mentioned in the comments, there are other reasons to call the eigenvalue problem $Ax=\lambda\cdot x$ nonlinear. For example, if one uses the characteristic polynomial $\chi_A(t)=\det(t\cdot I_n-A)$ to solve for the eigenvalues of $A$, then one ends up factoring an $n$th degree polynomial, which is a nonlinear problem.

It's also worth noting that taking any linear combination of two eigenvectors $x_1$ and $x_2$ corresponding to the same eigenvalue $\lambda$ does yield a solution to an eigenvalue problem. This is precisely the statement that eigenvectors corresponding to an eigenvalue $\lambda$ are organized into the eigenspace $E_\lambda=\operatorname{Null}(\lambda\cdot I_n-A)$.

Related Question