Newtonian Mechanics 04
— 1. Variational Calculus —
Definition 1 A functional is a mapping from vector spaces to into real numbers. |
Let $ {\displaystyle J = \int _{x_1}^{x^2} f\{y(x),y\prime (x),x\}dx}$. Suppose that $ {x_1}$ and $ {x_2}$ are constants, the functional form of $ {f}$ is known.
According to definition 1 $ {J}$ is a functional and the goal of the Calculus of Variations is to determine $ {y(x)}$ such that the value of $ {J}$ is an extremum.
Let $ {y=y(\alpha, x)}$ be a parametric representation of $ {y}$ such that $ {y(0,x)=y(x)}$ is the function that makes $ {J}$ an extremum.
We can write $ {y(\alpha, x)=y(0,x)+ \alpha\eta(x}$, where $ {\eta (x)}$ is a function of $ {x}$ of the class $ {C^1}$ (that means that $ {\eta}$ is a continuous function whose derivative is also continuous) with $ {\eta (x_1)=\eta (x_2)=0}$.
Now $ {J}$ is of the form $ {\displaystyle J(\alpha) = \int _{x_1}^{x^2} f\{y(\alpha, x),y\prime (\alpha, x),x\}dx}$. Therefore the condition for $ {J}$ to be an extremum is
$ \displaystyle \frac{dJ}{d\alpha}(\alpha=0)=0$
— 2. Euler Equations —
$ {\begin{aligned} \frac{\partial J}{\partial \alpha} &= \frac{\partial}{\partial \alpha} \int _{x_1}^{x_2}f(y,y\prime,x)dx \\ &= \int _{x_1}^{x_2}\left(\frac{\partial f}{\partial y}\frac{\partial y}{\partial \alpha}+ \frac{\partial f}{\partial y\prime}\frac{\partial y\prime}{\partial \alpha}\right) dx \end{aligned}}$
Since it is $ {\partial y /\partial \alpha = \eta (x)}$ and $ {\partial y\prime /\partial \alpha = d\eta/dx}$ it follows
$ \displaystyle \frac{\partial J}{\partial \alpha}= \int _{x_1}^{x_2}\left(\frac{\partial f}{\partial y}\eta (x)+ \frac{\partial f}{\partial y\prime}\frac{d \eta}{dx}\right) dx$
Now $ {\displaystyle \int _{x_1}^{x_2}\frac{\partial f}{\partial y\prime}\frac{d \eta}{dx}dx=\frac{\partial f}{\partial y\prime}\eta (x)|_{x_1}^{x_2}- \int _{x_1}^{x_2}\frac{d}{dx}\left( \frac{\partial f}{\partial y\prime} \right)\eta (x) dx}$.
For the first term it is $ {\frac{\partial f}{\partial y\prime}\eta (x)|_{x_1}^{x_2}=0}$ since $ {\eta (x_1)=\eta (x_2)=0}$ by hypothesis.
Hence
$ {\begin{aligned} \frac{\partial J}{\partial \alpha} &= \int _{x_1}^{x_2}\left(\frac{\partial f}{\partial y}\frac{\partial y}{\partial \alpha}- \frac{d}{dx}\left( \frac{\partial f}{\partial y\prime} \right) \frac{\partial y}{\partial \alpha}\right)dx \\ &= \int _{x_1}^{x_2}\left( \frac{\partial f}{\partial y}-\frac{d}{dx}\frac{\partial f}{\partial y\prime} \right)\eta (x) dx \end{aligned}}$
Remembering that $ {\partial J / \partial\alpha(\alpha=0)=0}$ and taking into account the fact that $ {\eta (x)}$ is an arbitrary function one can conclude that
$ \displaystyle \frac{\partial f}{\partial y}-\frac{d}{dx}\frac{\partial f}{\partial y\prime}=0 $
The previous equation is known as the Euler's Equation
To close our thoughts on the Euler equation let us say that there also is a second form for the Euler equation. The second form is
$ \displaystyle f-y\prime\frac{\partial f}{\partial y\prime}= \mathrm{const} $
and is used in the cases where $ {f}$ doesn't depend explicitly on $ {x}$.
— 3. Euler Equation for $ {n}$ variables —
Let $ {f}$ be of the form $ {f=f\{ y_1(x),y\prime _1(x),y_2(x),y\prime _2(x),\cdots,y_n(x),y\prime _n(x), x \}}$.
Now we have $ {y_i(\alpha, x)= y_i(0,x)+\alpha \eta (x)}$ and $ {\displaystyle \int _{x_1}^{x_2}\left( \frac{\partial f}{\partial y_i}-\frac{d}{dx}\frac{\partial f}{\partial y _i\prime} \right)\eta _i (x) dx}$ for each of the values of $ {i}$. Since $ {\eta _i(x)}$ are independent functions it follows that for $ {\alpha=0}$
$ \displaystyle \frac{\partial f}{\partial y_i}-\frac{d}{dx}\frac{\partial f}{\partial y _i\prime}=0 $
That is to say we have $ {n}$ independent Euler equations.
0 comments:
Post a Comment