That reduced matrix shows you that the set of vectors is linearly dependent for every value of $h$. If $h\ne 10$, the system has no solution, and if $h=10$, it has infinitely many, so there is no value of $h$ that gives it exactly one solution.
Indeed, you can see this directly from the vectors themselves: $v_2=-3v_1$.
It is true that two vectors are dependent if they "point in the same (or opposite) direction", i.e. if they are aligned.
But that is not totally true for three vectors in $3$D or more.
In the sense that, when the three vectors are aligned, i.e. parallel, i.e. when they are scalar multiples of each other, they are for sure dependent.
But the definition of linear dependency of three vectors is wider than being parallel: it includes also the case in which they are co-planar, although not parallel.
If you want to see that geometrically, taking the three vectors as position vectors from the origin, if they define a full $3$D parallelepiped then they are independent, if instead the parallelepiped collapses into a flat figure or segment then the vectors are dependent.
Algebraically this translates into the fact whether the matrix formed by the three vectors has full rank ($3$) or less.
Similarly for $n$ vectors of $m$ dimensions.
Then from the theory of linear system you know that, in a homogeneous system, if the matrix has full rank then it has the only solution $(0,0, \cdots, 0)$ which corresponds to the combination coefficients to be all null.
In reply to your comment, in ${\mathbb R}^2$ if you have two non-aligned = independent vectors, then a third one will lie on their same plane (the $x,y$ plane).
In the geometric interpretation, the parallelepiped (the hull) will be flat, i.e. dimension 2, which is less than 3, the number of vectors.
In the algebraic interpretation, a matrix $3 \times 2$ cannot have a rank greater than two: so 3 (or more) 2D vectors are necessarily dependent.
final note (to clarify what might be the source of your confusion)
The (in)dependence of $n$ vectors in ${\mathbb R}^m$ is defined for the whole set of $n$ vectors: they might be dependent, notwithstanding that a few of them ($q<n, \; q\le m$) could be independent. Yet if one is dependent on another (or other two, etc.), then the whole set is dependent.
And in fact it is a common task, given $n$ vectors, to find which among them represent an independent subset: the minor in the matrix with non-null determinant, the larger giving the rank.
Best Answer
Just build a matrix with these vectors as rows and perform row reduce. The vectors will be linearly dependent if at least one row is made of zeros. The idea is that the rank of a matrix is the maximum number of linearly independent rows (or columns), hence, the rows will be linearly dependent if and only if $r(A) < 3$.
$$ \begin{pmatrix} 1 & 4 & 1& -2 \\ -1 & a & b & 2\\ 1 & 1 & 1 & c\end{pmatrix}\to \begin{pmatrix} 1 & 4 & 1& -2 \\ 0 & a+4 & b+1 & 0\\ 0 & -3 & 0 & c+2\end{pmatrix}\to \begin{pmatrix} 1 & 4 & 1& -2 \\ 0 & -3 & 0 & c+2\\ 0 & a+4 & b+1 & 0\end{pmatrix} \to $$
$$ \begin{pmatrix} 1 & 4 & 1& -2 \\ 0 & -3 & 0 & c+2\\ 0 & 0 & b+1 & \frac{(c+2)(a+4)}{3}\end{pmatrix} $$
So the vectors are linearly dependent if and only if the last row is filled with zeros, i.e. $b = -1 \wedge (a=-4 \vee c=-2)$.