logo

Lagrangians and Euler-Lagrange Equations in Partial Differential Equations 📂Partial Differential Equations

Lagrangians and Euler-Lagrange Equations in Partial Differential Equations

Definition1

  • Lagrangian

    Let’s assume a smooth function L:Rn×RnRL : \mathbb{R}^n \times \mathbb{R}^n \to \mathbb{R} is given. This is called the Lagrangian and is denoted as follows.

    L=L(v,x)=L(v1,,vn,x1,,xn)v,xRnDvL=(Lv1,,Lvn),DxL=(Lx1,,Lxn) L = L(v,x)=L(v_{1}, \dots, v_{n}, x_{1}, \dots, x_{n}) \quad v,x\in \mathbb{R}^{n} \\ D_{v}L = (L_{v_{1}}, \dots, L_{v_{n}}), \quad D_{x}L = (L_{x_{1}}, \dots, L_{x_{n}})

    The reason for using variables v,xv, x is because, in physics, each variable actually signifies velocity and position.

  • Action, Admissible class

    For two fixed points x,yRnx,y \in \mathbb{R}^{n} and time t>0t>0, the following defined functional II is called action.

    I[w()]:=0tL(w˙(s),w(s))ds( ˙=dds) I[ \mathbf{w}(\cdot)] := \int_{0}^tL(\dot{\mathbf{w}}(s), \mathbf{w}(s) ) ds \quad \left( \dot{ }=\dfrac{d}{ds}\right)

    In this case, the function w()=(w1(),,wn())\mathbf{w}(\cdot)=\big( w^1(\cdot), \cdots, w^n(\cdot) \big) is an element of the admissible class defined as follows A\mathcal{A}.

    A:={w()=C2([0,t];Rn)  w(0)=y,w(t)=x} \mathcal{A} := \left\{ \mathbf{w}(\cdot) = \in C^2 \big([0,t];\mathbb{R}^n \big) \ \big| \ \mathbf{w}(0)=y, \mathbf{w}(t)=x\right\}

    In other words, A\mathcal{A} means the set of all paths that are twice continuously differentiable, starting from position yy to xx as time flows from 00 to tt.

Description

The goal of the calculus of variations is to find xA\mathbf{x} \in \mathcal{A} that makes the integral value of action II minimum. This x\mathbf{x} is called the minimizer of II.

I[x()]=infw()AI[w()] I[ \mathbf{x} (\cdot) ] = \inf_{\mathbf{w}(\cdot)\in \mathcal{A}} I[\mathbf{w}(\cdot)]

Such x\mathbf{x} is sought because the path that minimizes the action of the Lagrangian is actually the path along which an object moves. In essence, this is for understanding the motion of an object, which fundamentally equates to solving F=maF=ma. In classical mechanics, the Lagrangian specifically refers to the difference between kinetic energy and potential energy.

Regarding the determination of minimizers, there’s the following theorem.

Theorem

Assume x()A\mathbf{x}(\cdot) \in \mathcal{A} is the minimizer of action II. Then, x()\mathbf{x}(\cdot) satisfies the following equation.

dds[DvL(x˙(s),x(s))]+DxL(x˙(s),x(s))=0(0st) -\dfrac{d}{ds} \Big[ D_{v}L\big( \dot{\mathbf{x}}(s), \mathbf{x}(s) \big) \Big] + D_{x}L\big( \dot{\mathbf{x}}(s), \mathbf{x}(s)\big)=0 \quad (0 \le s \le t)

This equation is called the Euler-Lagrange equations.


It’s important to note that although a minimizer satisfies the Euler-Lagrange equations, satisfying the Euler-Lagrange equations does not necessarily mean it’s a minimizer. Similar to how differentiating at the minimum value gives 00, but not every point where differentiation gives 00 is a minimum. In this sense, x()A\mathbf{x}(\cdot) \in \mathcal{A} satisfying the Euler-Lagrange equations is called a critical point of II. Therefore, while a minimizer is a critical point, not all critical points are minimizers.

Proof

Assume xA\mathbf{x} \in \mathcal{A} is the minimizer of action II.

  • Step 1.

    Let’s presume the function y:[0,t]Rn,y()=(y1(),,yn())\mathbf{y} : [0,t] \to \mathbb{R}^{n}, \mathbf{y}(\cdot) = (y^1(\cdot), \cdots, y^n(\cdot) ) satisfies the following equation as a smooth function.

    y(0)=y(t)=0 \begin{equation} \mathbf{y}(0)=\mathbf{y}(t)=\mathbf{0} \label{eq1} \end{equation}

    And for any τR\tau \in \mathbb{R}, define w()\mathbf{w}(\cdot) as follows.

    w():=x()+τy()A \mathbf{w}(\cdot) : = \mathbf{x}(\cdot) + \tau \mathbf{y}(\cdot) \in \mathcal{A}

    Then w\mathbf{w} represents a path that has the same start and end points as x\mathbf{x} but differs by τy()\tau \mathbf{y}(\cdot) in between. Furthermore, since x()\mathbf{x}(\cdot) is the minimizer of II, the following equation holds.

    I[x()]I[w()]=I[x()+τy()]=:i(τ) I[\mathbf{x}(\cdot)] \le I[\mathbf{w}(\cdot)]=I[\mathbf{x}(\cdot) + \tau \mathbf{y}(\cdot)] =: i(\tau)

    Also, the function ii has a minimum value at τ=0\tau=0 by definition of a minimizer. Therefore, if the derivative of ii exists, it is i(0)=0i^{\prime}(0)=0.

  • Step 2.

    As defined above, ii is as follows.

    i(τ)=0tL(x˙(s)+τy˙(s),x(s)+τy(s))ds i(\tau) = \int_{0} ^t L\big( \dot{\mathbf{x}}(s) + \tau \dot{\mathbf{y}}(s), \mathbf{x}(s)+ \tau \mathbf{y}(s) \big)ds

    Since LL, y\mathbf{y} are smooth functions, by differentiating ii, we get the following.

    i(τ)=0ti=1n[Lvi(x˙+τy˙,x+τy)y˙i+Lxi(x˙+τy˙,x+τy)yi]ds i^{\prime}(\tau) = \int_{0}^t \sum_{i=1}^{n} \left[ L_{v_{i}} ( \dot{\mathbf{x}} + \tau \dot{\mathbf{y}}, \mathbf{x}+ \tau \mathbf{y} )\dot{y}^i + L_{x_{i}} ( \dot{\mathbf{x}} + \tau \dot{\mathbf{y}}, \mathbf{x}+ \tau \mathbf{y} ) y^i \right] ds

    Substituting τ=0\tau=0, we obtain the following result from Step 1.

    0=i(0)=0ti=1n[Lvi(x˙,x)y˙i+Lxi(x˙,x)yi]ds 0=i^{\prime}(0) = \int_{0}^t \sum_{i=1}^{n} \left[ L_{v_{i}} ( \dot{\mathbf{x}} , \mathbf{x})\dot{y}^i + L_{x_{i}} ( \dot{\mathbf{x}} , \mathbf{x} )y^i \right] ds

    Applying partial integration to each term of Lviy˙iL_{v_{i}}\dot{y}^i, and by the assumption y(0)=y(t)=0\mathbf{y}(0)=\mathbf{y}(t)=\mathbf{0}, we get the following.

    0tLviy˙ids=Lviyi]0t0tddsLviyi=0tddsLviyi \int_{0}^t L_{v_{i}}\dot{y}^i ds = \left. L_{v_{i}}y^i \right]_{0}^t- \int_{0}^t \dfrac{d}{ds}L_{v_{i}}y^i=\int_{0}^t-\dfrac{d}{ds} L_{v_{i}}y^i

    Therefore, the following equation holds.

    0=i(0)=i=1n0t[ddsLvi(x˙,x) +Lxi(x˙,x)]yids 0=i^{\prime}(0) = \sum_{i=1}^{n} \int_{0}^t \left[ -\dfrac{d}{ds} L_{v_{i}} ( \dot{\mathbf{x}} , \mathbf{x})\ + L_{x_{i}} ( \dot{\mathbf{x}} , \mathbf{x} ) \right]y^i ds

The result of Step 2. satisfies for all smooth functions y:[0,t]Rn\mathbf{y} : [0,t] \to \mathbb{R}^n that fulfill (eq1)\eqref{eq1}. Therefore, the value inside the brackets must be 00. Thus, the following is true.

ddsLvi(x˙,x)+Lxi(x˙,x)=0 -\dfrac{d}{ds} L_{v_{i}}( \dot{\mathbf{x}}, \mathbf{x} ) +L_{x_{i}}( \dot{\mathbf{x}}, \mathbf{x}) =0


This result leads to the Hamiltonian equations.

See Also


  1. Lawrence C. Evans, Partial Differential Equations (2nd Edition, 2010), p115-117 ↩︎