2.3: General Method for the Minimization Problem
( \newcommand{\kernel}{\mathrm{null}\,}\)
To emphasize the generality of the method, we’ll just write
J[y]=∫x2x1f(y,y′)dx(y′=dy/dx)
Then under any infinitesimal variation δy(x) (equal to zero at the fixed endpoints)
δJ[y]=∫x2x1[∂f(y,y′)∂yδy(x)+∂f(y,y′)∂y′δy′(x)]dx=0
To make further progress, we write δy′=δ(dy/dx)=(d/dx)δy, then integrate the second term by parts, remembering δy=0 at the endpoints, to get
δJ[y]=∫x2x1[∂f(y,y′)∂y−ddx(∂f(y,y′)∂y′)]δy(x)dx=0
Since this is true for any infinitesimal variation, we can choose a variation which is only nonzero near one point in the interval, and deduce that
∂f(y,y′)∂y−ddx(∂f(y,y′)∂y′)=0
This general result is called the Euler-Lagrange equation. It’s very important—you’ll be seeing it again.