[Physics] Principle of Least Action via Finite-Difference Method

actionclassical-mechanicslagrangian-formalismvariational-principle

I am reading Gelfand's Calculus of Variations & mathematically everything makes sense to me, it makes perfect sense to me to set up the mathematics of extremization of functionals & show that in extremizing a certain functional you can end up with Newton's laws (i.e. you could extremize some arbitrary functional $L$ & in examining the form of Newton's laws you see that one should define $L$ as $T – V$) & this way of looking at things requires no magic, to me it seems as though you've just found a clever way of doing mathematics that results in Newton's laws.

However in books like Landau one must assume this magical principle of least action using the kind of thinking akin to Maupertuis, & I remember that every time I read some form of justification of this there's always a crux point where they'd say 'because it works'. I'm thinking that there may be a way to explain the principle of least action if you think of extremizing functionals along the lines Euler first did & use the method of finite differences (as is done in the chapter on the Variational Derivative in Gelfand, can't post a link unfortunately), i.e. because you're thinking of a functional as a function of n variables you can somehow incorporate the $T – V$ naturally into what you're extremizing, but I don't really know… I'm really hoping someone in here can give me a definitive answer, to kind of go through what the thought process is in this principle once and for all!

As far as I know, one can either assume this as a principle as in Landau & use all this beautiful theory of homogeneity etc… to get some results, or else once can assume Newton's laws & then use the principle of virtual work to derive the Euler-Lagrange equations, or start from Hamilton's equations & end up with Newton's laws, or one can assume Newton's laws & set up the calculus of variations & show how extremizing one kind of functional leads to Newton's laws – but I'm not clear on exactly what's going on & would really really love some help with this.

Best Answer

I) OP's questions about the variational principle (e.g. the derivation to and from Newton's laws) are good and interesting questions, but most have been asks before, see e.g. this and this Phys.SE posts and links therein. Lagrangians not of the form Kinetic Energy minus Potential Energy are discussed in this Phys.SE post. The question of whether (or not) a set of equations of motion has an action principle is a duplicate of this Phys.SE post and links therein. Another possible helpful Phys.SE post concerning calculus of variation is this one.

II) The only new ingredient in OP's question seems to be the finite-difference method. Gelfand's use of the finite-difference method seems mostly to be an alternative way to gain insight into the Euler-Lagrange formula for the functional derivative via discretization. It doesn't add anything new to the continuum theory.

For instance, let us for simplicity consider point mechanics (as opposed to field theory). A continuum formulation for point mechanics here refers to a continuous time variable $t$ (as opposed to a discrete time variable). Let there be given a continuum Lagrangian $L(q,v,t)$ that depends on three arguments: position $q$, velocity $v$ and time $t$. Let $\frac{\partial L}{\partial q}$ and $\frac{\partial L}{\partial v}$ denote the derivative wrt. the first and second argument, respectively, and so forth.

Let us approximate the continuum action functional

$$\tag{1} S[q]~:=~\int_{0}^{T} \!\! dt~L(q(t),\dot{q}(t),t)$$

via the following discretized expression

$$\tag{2} S(q_0, \ldots q_{N})~:=~\Delta t\sum_{n=0}^{N-1} L(q_n,v_n,t_n), $$

where

$$\tag{3} \Delta t~:=~ \frac{T}{N}, \qquad t_n ~:= n \Delta t, \qquad q_n~:=~q(t_n),\qquad v_n~:=~\frac{q_{n+1}-q_n}{\Delta t}.$$

In eq. (3) we have chosen the forward difference for $v_n$. Obviously other choices are possible. Then

$$\frac{1}{\Delta t}\frac{\partial S(q_0, \ldots q_{N})}{\partial q_n} $$ $$\tag{4}~=~ \frac{\partial L(q_n,v_n,t_n)}{\partial q} -\frac{1}{\Delta t}\left[\frac{\partial L(q_n,v_n,t_n)}{\partial v}-\frac{\partial L(q_{n-1},v_{n-1},t_{n-1})}{\partial v} \right], $$

where $n\in\{1,2, \ldots,N-2, N-1\}$. Eq. (4) is a discretized version of the Euler-Lagrange formula

$$\tag{5} \frac{\delta S[q]}{\delta q (t)}~=~\frac{\partial L(q(t),\dot{q}(t),t)}{\partial q(t)} - \frac{d}{dt} \frac{\partial L(q(t),\dot{q}(t),t)}{\partial \dot{q}(t)} $$

for the functional derivative$^1$. The functional derivative (5) satisfies by definition the following infinitesimal variation formula

$$\tag{6} \delta S[q] ~:=~S[q+\delta q]- S[q]~=~ \int_{0}^{T} \!\! dt~\frac{\delta S[q]}{\delta q (t)}\delta q(t). $$

--

$^1$ The existence of functional derivative depends on the appropriately imposed boundary conditions, which we have not discussed. In plain English: We need to integrate by part at some point.

Related Question