[Physics] Why do we need to renormalize in QFT, really

quantum-field-theoryrenormalization

What is the modern understanding of why UV infinities arise in Quantum Field Theories? While nowadays there is clearly a consensus about how to remove them, I'm interested in the reason why they show up in the first place.

Below are a few thoughts of mine but please feel free to ignore them. And please ignore the so-far only answer by Lubos Motl.


So… UV divergences arise and thus we need to renormalize, because:

  1. We have infinite number of degrees of freedom ín a field theory. (From this perspective, the infinites seem inevitable.)
  2. We multiply fields to describe interactions, fields are distributions and the product of distributions is ill-defined.
  3. We neglect the detailed short-wavelength structure of scattering processes, and the infinites are a result of our approximations with delta potentials. (From this point of view, the UV divergences aren't something fundamental, but merely a result of our approximation method. )
  4. We are dealing with non-self-adjoint Hamiltonians. (This is closely related to the 3. bullet point. From this perspective an alternative to the "awkward" renormalization procedure would be the "method of self-adjoint extension".)

Are these reasons different sides of the same coin? And if yes, how can we understand the connection between them?


Edit: The "answer" by Luboš Motl is not an answer at all. He simply states: Yeah there are infiniteies, but you can get rid of them, so you shouldn't care about them. Basically "Stop trying to understand things; shut up and calculate". Everyone is free to take this attitude, but it certainly doesn't explain why we encounter infinities in the first place and hence we the renormalization procedure is needed. Because his answers gets a surprisingly large number of upvotes and no one else want's to write another answer, because Motl's answer appears as the "accepted answer", I feel the need to add some more thoughts here:

In contrast, a much more interesting perspective is mentioned by Batterman in "Emergence, Singularities, and Symmetry Breaking":

Many physicists and philosophers apparently believe that singularities appearing in our theories are indications of modeling failures. In particular, they hold that such blowups are signatures of some kind of unphysical assumption in the underlying model. Singularities are, on this view, information sinks. We cannot learn anything about the physical system until we rid the theory of such monstrosities. (This attitude is still quite prevalent in the literature on quantum field theory, a little bit of which will be discussed in the next section.) On the contrary, I’m suggesting that an important lesson from the renormalization group successes is that we rethink the use of models in physics. If we include mathematical features as essential parts of physical modeling then we will see that blowups or singularities are often sources of information. Mark Wilson has noted something akin to this in his discussion of determinism in classical physics. He notes that while from a modeling point of view we are inclined to object to the appearance of singularities . . . , from a mathematical point of view we often greatly value these same breakdowns; for as Riemann and Cauchy demonstrated long ago, the singularities of a problem commonly represent the precise features of the mathematical landscape we should seek in our efforts to understand how the qualitative mathematics of a set of equations unfolds. Insofar as the project of achieving mathe-matical understanding goes, singularities frequently prove our best friends, not our enemies [24, pp. 184–185].

and in a similar spirit here's a quote from Roman Jackiew:

“the divergences of quantum field theory must not be viewed as unmitigated defects; on the contrary they convey crucially important information about the physical situation, without which most of our theories would not be physically acceptable.”


Moreover, some background on the origin of my question: Some of these reasons are mentioned in the form of questions at page 6 in "An introduction to the nonperturbative renormalization group" by Bertrand Delamotte. However, after stating the questions he only mentions that

"Some of these questions will be answered directly in the following. For some others the reader will have to build his own answer. I hope that these notes will help him finding answers."

Unfortunately, I wasn't able to find conclusive answers in his paper.


I found a nice discussion of "The mystery of renormalization" in "Phase Transitions and Renormalization Group" by Zinn-Justin:

Though it was now obvious that QED was the correct theory to describe electromagnetic interactions, the renormalization procedure itself, allowing the extraction of finite results from initial infinite quantities, had remained a matter of some concern for theorists: the meaning of the renormalization ‘recipe’ and, thus, of the bare parameters remained obscure. Much effort was devoted to try to overcome this initial conceptual weakness of the theory. Several types of solutions were proposed:

(i) The problem came from the use of an unjustified perturbative expansion and a correct summation of the expansion in powers of α would solve it. Somewhat related, in spirit, was the development of the so-called axiomatic QFT, which tried to derive rigorous, non-perturbative, results from the general principles on which QFT was based.

(ii) The principle of QFT had to be modified: only renormalized perturbation theory was meaningful. The initial bare theory with its bare parameters had no physical meaning. This line of thought led to the BPHZ (Bogoliubov, Parasiuk, Hepp, Zimmerman) formalism and, finally, to the work of Epstein and Glaser, where the problem of divergences in position space (instead of momentum space) was reduced to a mathematical problem of a correct definition of singular products of distributions. The corresponding efforts much clarified the principles of perturbation theory, but disguised the problem of divergences in such a way that it seemed never having existed in the first place.

(iii) Finally, the cut-off had a physical meaning and was generated by additional interactions, non-describable by QFT. In the 1960s some physicists thought that strong interactions could play this role (the cut-off then being provided by the range of nuclear forces). Renormalizable theories could then be thought as theories some- what insensitive to this additional unknown short-distance structure, a property that obviously required some better understanding.

This latter point of view is in some sense close to our modern understanding,
even though the cut-off is no longer provided by strong interactions.


Best Answer

None of the four explanations is quite right.

  1. We know theories with finite observables but an infinite number of degrees of freedom
  2. the delta-functions and distributions appear already at the tree level but the divergences that need to be renormalized only appear in loops, so they have nothing to do with each other
  3. it is really equivalent to (2)
  4. Hamiltonians in QFTs are Hermitian

Instead, there are old-fashioned correct reasons to explain why renormalization is needed, and the modern explanation based on the renormalization group paradigm.

The old-fashioned need for renormalization is simple: when one calculates the loop corrections to Green's functions or scattering amplitudes, and assumes that the bare coupling constants in the Lagrangian are finite, then he gets infinite contributions from the integrals over very high (or very low, IR divergences) momenta of virtual particles. That means that the total result is infinite as well which can't match finite observations. So the solution is to assume that the bare constants aren't really finite, they have infinite terms in them as well, and the infinities cancel between those infinite parts of the coupling constants (or between counterterms) on one side and the loop divergences on the other.

The modern explanation why renormalization is needed is the renormalization group paradigm. It says that all quantum field theories we use are really "effective", i.e. good for answering questions about low-energy enough processes. It means that one must really clarify the theory and express the "typical energy scale" where we still want the theory to be valid. When this dependence on the scale is studied carefully, we find out that the dimensionless constants such as the fine-structure constant aren't really dimensionless, or aren't really constant. They run – depend on the energy – in a slower way and this running is the reason why the limit is singular if we want the theory to work at arbitrarily high energies or short distance scales.

I would recommend you to focus on the question how the quantities are actually calculated, and not just why they need infinite subtractions. Once you understand how a calculation actually proceeds, you will know about a more useful and important thing – how to calculate – and you will also understand why your guesses about the "causes of divergences and renormalization" differ from the real ones.

Related Question