Your proof is valid for a bounded function defined on the closed, bounded interval $[a,b]$, despite the apparent simplicity. It also relies on the fact that the Riemann and Lebesgue integrals are the same for step functions, as mentioned by Tony Piccolo.
The other proof you mention is also valid but takes you on a more roundabout path because it strings together a number of results, each of which is not altogether trivial to prove.
A bounded, measurable function defined on a set of finite
measure is Lebesgue integrable.
If a sequence of measurable functions converges almost everywhere
to $f$, then the limit function $f$ is measurable.
If $f$ is Riemann integrable on $[a,b]$, then there exists a
sequence of simple (measurable) functions converging almost everywhere
to $f$.
Adding even more complexity, the proof of the third statement that I know also uses the fact that the set of discontinuities of a Riemann integrable function must be of measure zero. It begins by constructing a partition of $[a,b]$ with dyadic intervals:
$$I_{n,k} = \begin{cases}[a + (b-a)\frac{k-1}{2^n}, a + (b-a)\frac{k}{2^n})\,\,\, k = 1 , \ldots, 2^n -1 \\ [a + (b-a)\frac{2^n-1}{2^n}, b] \,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\, k = 2^n\end{cases} $$
With $m_k = \inf_{I_{n,k}}f(x)$, we can construct the sequence of simple functions
$$\phi_n(x) = \sum_{k=1}^{2^n} m_k \, \chi_{I_{n,k}}(x)$$
The sequence is increasing and, since $f$ is bounded, by monotonicity is convergent to some function $\phi$ such that $\phi_n(x) \uparrow \phi(x) \leqslant f(x).$ If $x \in I_{n,k}$ then the oscillation of $f$ over that interval satisfies
$$\sup_{u,v \in I_{n,k}}|f(u) - f(v)| \geqslant f(x) - \phi_n(x) \geqslant f(x) - \phi(x).$$
Thus, $f(x) \neq \phi(x)$ only at points where $f$ is not continuous, which must belong to a measure zero set if $f$ is Riemann integrable. Therefore, $f$ is almost everywhere the limit of a sequence of simple functions and is measurable.
I would agree that the second approach is "pretty complicated". However, I think it is not uncommon to find theorems with a variety of proofs ranging in complexity.
Best Answer
Set $f^+=\max\{f,0\}$ and $f^-=\max\{-f,0\}$. A function is integrable if and only if the positive part $f^+$ and the negative part $f^-$, which are both nonnegative, are integrable meaning $$0\leq\int f^+ dx<\infty \text{ and } 0\leq\int f^- dx<\infty.$$
Now, check that by hypothesis, $g\leq h$ implies $g^+\leq h^+$ and so if $\int h^+ dx<\infty$, then $\int g^+ dx<\infty$ by comparison theorem. In the same way, $f\leq g$ implies $g^-\leq f^-$ and integrability of $f^-$ implies that of $g^-$. So $g$ is integrable.
Now that the integrability is proved, you just take the integral of all sides and if the first integral equals the third integral, the second integral also must be equal to the other two (circle inequalities in algebra: $a\leq b\leq a$ implies $a=b$)