I happened to be thinking recently about how to tell when a functional is extremized. Examples in physics include minimizing the ground state energy of an electronic system expressed as an approximate density functional E[ρ] with respect to the electron density ρ or maximizing the relativistic proper time τ of a classical particle with respect to a path through spacetime. Additionally, finding the points of stationary action that lead to the Euler-Lagrange equations of motion is often called "minimization of the action", but I can't recall ever having seen a proof that the action is truly minimized (as opposed to reaching a saddle point). This got me to think more about the conditions under which a functional is truly maximized or minimized as opposed to reaching a saddle point. Follow the jump to see more. I will frequently refer to concepts presented in a recent post (link here), including the relationships between functionals of vectors & functionals of functions. Additionally, for simplicity, all variables and functions will be real-valued.
Determining the Nature of a Stationary Point
For a function f of a single variable x, the nature of a stationary point can be determined as follows. If x0 is a stationary point such that dfdx|x0=0, then the condition for it to correspond to a minimum (maximum) is that d2fdx2|x0 is positive (negative); if the second derivative vanishes, then it may be a saddle point, though this may not be the case if it turns out that there exists a nonvanishing higher-order even derivative. (An example is f(x)=x4, which clearly has a minimum at x0=0 but which has d2fdx2|x0=0, so one must use d3fdx3|x0=0 and d4fdx4|x0>0 to make the case.)
For a functional F of multiple variables, which may be collected as a vector v, the nature of a stationary point can be determined as follows. If w is a stationary point such that ∂F∂vi|w=0 for every index i, then the condition for it to correspond to a minimum (maximum) is that the [symmetric] Hessian matrix whose elements are given by ∂2F∂vi∂vj|w is positive definite (negative definite). If the matrix has some eigenvalues that are positive and others that are negative, then the stationary point is a saddle point. If the matrix has nonvanishing eigenvalues that are all the same sign but also has vanishing eigenvalues, then higher-order derivatives must be used to determine the nature of the stationary point, though writing this out explicitly is often much more cumbersome especially as the derivative of order n produces a "matrix" (perhaps more properly a tensor) with n indices.
This suggests that for a functional F of a function g, the nature of a stationary point can be determined as follows. If f is a stationary point such that δFδg(x)|f=0 for every continuous index x, then the condition for it to correspond to a minimum (maximum) is that the [symmetric] Hessian operator whose elements are given in position space by δ2Fδg(x)δg(x′)|f is positive definite (negative definite). In particular, using the definition K(x,x′)=δ2Fδg(x)δg(x′)|f, it is possible to perform a spectral decomposition into eigenvalues λ(θ) and eigenfunctions u(x,θ) (assumed to be indexed continuously by a variable \theta as the functions are assumed to be supported over the entire real line) such that K(x,x′)=∫λ(θ)u(x,θ)u(x′,θ) dθ. Thus, if λ(θ)>0 for all θ, then the functional is minimized, while if λ(θ)<0 for all θ, then the functional is maximized. If λ(θ) changes sign, then the functional has reached a saddle point, while if λ(θ) vanishes for some θ but otherwise never changes sign, then higher-order derivatives must be considered, presenting the same difficulties as for functionals of vectors.
How Likely is a Functional to Have a True Maximum or Minimum?
As I thought about it more, I recognized that as the dimension of a vector space grows, it seems unlikely that a stationary point of a functional would be a true maximum or minimum, as the addition of new dimensions would open new opportunities for flipping the sign of a new eigenvalue of the Hessian matrix. In the case where the dimensions correspond to physical spatial dimensions, there would have to be clear physical constraints specific to the problem that would allow one to argue that the directions along which an eigenvalue of the Hessian matrix could be negative can be neglected. However, as the number of dimensions increases without bound (which is equivalent to considering a functional of functions instead of a functional of finite-dimensional vectors, as both countably & uncountably infinite sets of basis vectors can be used to represent infinite-dimensional vector spaces), it becomes less common for these dimensions to represent physical spatial dimensions and more common for these dimensions to represent other quantities, like frequencies, energies, momenta/wavevectors, or other things like that. (This statement is meant to imply only correlation, not causation in any direction of the statement.) In such cases, it may be easier to argue on general physical grounds for the neglect of negative eigenvalues of the Hessian matrix if they correspond to unphysically high energies, momenta, frequencies, or things like that. With such neglect, it may be easier to then argue that the stationary point of the functional is, to the degree that the aforementioned approximation is justified, a true maximum or minimum.
Second Derivative of the Nonrelativistic Classical Action for a Single Degree of Freedom
The nonrelativistic classical action for a single degree of freedom is typically given by S[x;t]=∫t0L(t′,x,˙x) dt′
To derive the second derivative, it may be better to define a new functional J[x;t]=δSδx(t) that depends on the trajectory x and is parameterized by t. This means δ2Sδx(t)δx(t′)=δJ[x;t]∂x(t′). From above, it can be seen that J[x;t]=∫∞−∞f(t,t″,x,˙x) dt″
In general, this may be quite nasty to expand. This is because total time derivatives appear in various places, and the existence of the Dirac delta function without an integral means that switching the position of the total time derivative operator becomes quite tricky.
Second Derivative of a Typical Newtonian Action for a Single Degree of Freedom
Things become much simpler in the case that L(t,x,˙x)=m˙x22−V(x)
The second derivative can be justified as follows, using the analogy to functionals of finite-dimensional vectors. If the function x(t) is replaced by the vector ui with discrete indices i and the derivative operator ddt is replaced by the antisymmetric matrix with elements Dij, then the first variation is ∂S∂ui=−∂V∂x|ui−m∑k,lDikDklul
It may be worth noting that the continuum expression looks sort of like a nonrelativistic single-particle quantum Hamiltonian acting on the wavefunction of a particle localized at a single point in space x0, which in position space looks like V(x)δ(x−x0)−ℏ22m∂2∂x2δ(x−x0). When going from the quantum case to the classical case, the role of position x is replaced by time t, the role of the wavefunction is replaced by an identity operator δ(t−t′) in the time domain, the potential energy V(x) is replaced by its negative second derivative −∂2V∂x2 evaluated along the trajectory x(t), and the kinetic energy of the wavefunction in position space given by the operator −ℏ22m∂2∂x2 is replaced by the acceleration term −md2dt2 (which ultimately comes from the same kinetic energy in the classical Lagrangian or Hamiltonian). However, I would be wary of trying to ascribe any deeper meaning to this analogy, though it could be fruitful in showing how quantum & classical intuitions can overlap.
Returning to the larger point, the second derivative of the action is essentially a linear operator −∂2V∂x2|x(t)−md2dt2 when evaluated for a given trajectory x(t). The negative second derivative operator with respect to time is positive-definite. However, the effective "spring constant" given the trajectory −∂2V∂x2|x(t) might not always be negative, and depending on the competition between it and the negative second derivative operator with respect to time, the overall operator might be indefinite, meaning that the corresponding trajectory is a saddle point. The following few sections will discuss specific solvable examples.
Particle Experiencing a Uniform Force
Given the coordinate x and the potential V(x)=−F0x
Particle Experiencing a Harmonic Force
The potential for a particle experiencing both a harmonic restoring force and an external harmonic drive can be written as V(x)=k2x2−F0xcos(ωDt)
An Analogy to the Quantum Harmonic Oscillator
It might be interesting to try to reverse-engineer a potential such that the second derivative of the action gives an operator with the same structure as that of a quantum harmonic oscillator. In particular, this means constructing a potential such that δ2Sδx(t)δx(t′)=(kω20t2−md2dt2)δ(t−t′)
Particle in an Exponential Potential
It might be more instructive to consider a potential which yields analytically solvable equations of motion while also yielding an action whose second derivative depends nontrivially on the trajectory. An example of this could be the exponential potential V(x)=V0exp(x/x0)