The Heisenberg inequalities reads :
$$\Delta x \Delta p_x \geq \frac{\hbar}{2},$$
where
$$\hbar = \frac{h}{2\pi}. $$
Therefore, for a free particle, your add the expressions
$$ p_x=mv_x$$
so you get :
$$\Delta x \Delta v_x \geq \frac {\hbar}{2m}$$
you have the uncertainty on the velocity : $\Delta v_x \geq \frac{\hbar}{2m\Delta x}$
the Heisenberg inequalities relies two observables. here, $x$ and $p_x$ or $x$ and $v_x$.
Be carefull, if the particule is not free, there might be some changes.
I would boldly claim that this thought experiment (also known as the Heisenberg microscope) is simply the wrong picture to understand the origin of uncertainty principle. The reason why it is so is because it mixes up between uncertainty due to measurement and uncertainty due to quantum state; nonetheless it had made its way into numerous textbooks and confused numerous undergraduates (including me) by including quantum mechanical objects such as electrons and photons and giving some results that has the factor $\hbar$ in it.
I will try to explain this confusing business to the best of my abilities about your questions in three parts - firstly, what is Heisenberg uncertainty principle; secondly, why is it unique to quantum mechanics; and finally, why the Heisenberg microscope is a wrong way of understanding the uncertainty principle. I am sorry that I may have to include a bit of maths from time to time, but I hope you will follow (and I hope I am right about this - do comment if I made mistakes).
Firstly, what is uncertainty principle? The best way that I know of to understand it physically is the following scenario: imagine that you have prepared a huge quantity of identical quantum states, and you measured the position of half of these states and the momentum of the rest with perfect precision (see below). At the end of the day, you will obtain a list of positions and momenta, you will notice that these results do have uncertainties due to the probabilistic nature of quantum states.
Here is where the uncertainty principle kicks in: regardless of what quantum state you prepared in the first place, if you calculate the uncertainties of positions and momenta respectively by the data you obtained from that long list, it will always be the case the uncertainties calculated from the list obey the uncertainty principle $\Delta x \Delta p \geq \hbar/2$. A more interesting way of rephrasing it would be you can never prepare a quantum state of which the uncertainties calculated from the list $\Delta x \Delta p$ is smaller than $\hbar/2$.
Before moving on, it is worth discussing a few things in this imaginary scenario. First thing is obviously what do I mean by the phrase with perfect precision? I certainly do not mean that there is some 'position' and 'momentum' that the quantum state has prior to measurement, what I meant is that the measurement results are completely due to the quantum states themselves, and are subjected to no external disturbance by other physical systems. Well you may argue that it is physically impossible to do that for any experimental apparatus would introduce some perturbation of the system, but since we are living in the imaginary thought experiment well we get to decide what we can do and what we can't do.
And here's the point which is very important under the context of the problem: even in this ideal world we can obtain positions and momenta directly from the quantum states, the uncertainty principle still holds. Throw away apparatus like the microscope or any other fancy equipments, you still have uncertainty - and this property is fundamentally due to the nature of quantum states themselves.
Still need convincing why this is justified? Well here we enter the second part on uncertainty due to measurements. Look back to any experiments with classical systems - you can almost certainly find no experiments where there is 0% uncertainty as there are bound to be errors introduced by the environment; nonetheless it doesn't stop us from imagining a perfect experiment where the results are completely due to the physical system we are studying. Say you are measuring acceleration due to gravity in a lab - you can be certain that almost nobody will ever get $9.80665$ meter per second squared (unless you are a cheater) because of errors due to gravitational attraction to the surrounding objects, the grids on your ruler are not fine enough, etc. etc. But you have no problem convincing yourself that under the perfect and ideal condition you still will be able to get $9.80665$.
And the crux of the matter here is that the uncertainty due to environment (or errors) happens to all systems, be it classical or quantum. Nonetheless, the uncertainty principle only applies to quantum systems. In Newtonian mechanics, you can characterise the motion of a particle in one dimension by a pair of quantities $(x,p)$, or position and momentum, and you can make it such that following the experimental procedure we described in part 1, that by preparing a huge number of identical states and measure their positions and momenta, $\Delta x \Delta p \leq \hbar/2$. In fact, it is very easy: by preparing a bunch of particles having same position and momentum, $\Delta x = \Delta p = 0$. But in quantum mechanics, it simply cannot be done, because we are talking about an entirely different beast here: instead of $(x,p)$, you need to describe a quantum state with a wavefunction $|\psi\rangle$, and they must obey the uncertainty principle.
So here we are at the third part - why is the Heisenberg microscope the wrong picture to understand the origin of uncertainty principle. I suspect that you can now already answer that - the thought experiment basically attributes the origin of the uncertainty principle to error introduced in the experiment, but not the quantum state itself. In a perfect experiment, according to Heisenberg microscope, there will be no uncertainty; we can even try to perceive measuring the position and momentum of the electron using other methods - say shooting one electron off a gun and bouncing them off by a wall (maybe?) - that can give you uncertainties below the uncertainty principle according to the picture described by Heisenberg microscope. But this is simply not the case and you simply can't do that - because the state is described by a wavefunction $|\psi \rangle$, but not a pair of $(x,p)$, so it is simply wrong to use '$x$' or '$p$' to describe the electron.
This also leads to the complication about interactions, as you have mentioned in your question. The interaction between photon and electron cannot be simply described by 'momentum transfer' for this implicitly assumes that the physical state photons and electrons are characterised by some momenta. As stated before, the interaction can only be described in terms of $|\psi\rangle$; and to be absolutely strict the best way of understanding such interaction is from QED, rather than this semi-classical picture. Nonetheless, let me reiterate my statement again - remove the interaction (regardless of whether it is photon-electron interaction or whatever physical processes you use to probe the electron), you still have uncertainty principle, because it is a fundamental property of a quantum state.
Regardless, I suspect the reason why the Heisenberg microscope is so successful is because the way it mixes quantum mechanical interactions between electrons and photons and classical interpretations to give results involving the infamous $\hbar$ simply by manipulating the errors introduced in an experiment, which gives us the illusion that we can intuitively understand the uncertainty principle and it is simply not the case. I feel it's fitting to use this (mis)quote - certainly quantum mechanics has never allowed herself to be won; and at present every kind of intuition stands with sad and discouraged mien—IF, indeed, it stands at all! - but this is, I guess, why we love quantum mechanics so much :)
Best Answer
Manishearth's answer is correct, and this is just a minor extension of it. Manishearth correctly points out that the problem is your statement:
Your statement is the hidden variables idea, and courtesy of Bell's theorem we currently believe that hidden variables are impossible.
Take the example of a hydrogen atom, and ask what the position of the electron is. The problem is that properties like position are properties of particles. It doesn't make sense to ask what the position is unless there is a particle at that position. But the electron is not a particle. The question of what an electron really is may entertain philosophers, but for our purposes it's an excitation in a quantum field and as such doesn't have a position. If you interact with the electron, e.g. by firing another particle at it, you will find that the interaction between the particle and electron happens at a well defined position. We tend to think of that as the position of the electron, but really it isn't: it's the position of the interaction.
The uncertainty principle applies because it's not possible for an interaction, like our example of a colliding particle, to simultaneous measure the position and momentum exactly. So you're sort of correct when you say it's an observational limit, but it's a fundamental one.