R5.1 Proof that exponentiation of Transverse of a Matrix equals the Transverse of the Exponentiation Expansion Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
A
∈
R
(
n
×
n
)
{\displaystyle \displaystyle \mathbf {A} \in \mathbb {R} ^{(n\times n)}}
and
exp
[
A
]
n
×
n
=
I
n
×
n
+
1
1
!
A
n
×
n
+
1
2
!
A
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
A
k
n
×
n
{\displaystyle {\underset {\color {red}n\times n}{\exp[\mathbf {A} ]}}={\underset {\color {red}n\times n}{\mathbf {I} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}\mathbf {A} ^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {A} ^{k}}}}
(Eq.(2)p.20-2b )
exp
A
T
=
(
exp
A
)
T
{\displaystyle \displaystyle \exp \mathbf {A} ^{T}=(\exp \mathbf {A} )^{T}}
(Eq.5.1.1 )
We will first expand the LHS, then the RHS of (Eq. 5.1.1 ) using (Eq.(2)p.20-2b ) and compare the two expressions.
Expanding the LHS,
exp
[
A
T
]
n
×
n
=
I
T
n
×
n
+
1
1
!
A
T
n
×
n
+
1
2
!
[
A
T
]
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
[
A
T
]
k
n
×
n
{\displaystyle {\underset {\color {red}n\times n}{\exp[\mathbf {A^{T}} ]}}={\underset {\color {red}n\times n}{\mathbf {I^{T}} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A^{T}} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}[\mathbf {A^{T}} ]^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {[} {A^{T}}]^{k}}}}
But we know that
I
T
=
I
{\displaystyle \mathbf {I} ^{T}=\mathbf {I} }
∴
exp
[
A
T
]
n
×
n
=
I
n
×
n
+
1
1
!
A
T
n
×
n
+
1
2
!
[
A
T
]
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
[
A
T
]
k
n
×
n
{\displaystyle \therefore {\underset {\color {red}n\times n}{\exp[\mathbf {A^{T}} ]}}={\underset {\color {red}n\times n}{\mathbf {I} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A^{T}} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}[\mathbf {A^{T}} ]^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {[} {A^{T}}]^{k}}}}
(Eq.5.1.2 )
Now expanding the RHS,
exp
[
A
]
T
n
×
n
=
[
I
n
×
n
+
1
1
!
A
n
×
n
+
1
2
!
[
A
]
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
[
A
]
k
n
×
n
]
T
{\displaystyle {\underset {\color {red}n\times n}{\exp[\mathbf {A} ]^{T}}}=\left[{\underset {\color {red}n\times n}{\mathbf {I} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}[\mathbf {A} ]^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {[} {A}]^{k}}}\right]^{T}}
Which on calculating, reduces to
exp
[
A
T
]
n
×
n
=
I
T
n
×
n
+
1
1
!
A
T
n
×
n
+
1
2
!
[
A
T
]
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
[
A
T
]
k
n
×
n
{\displaystyle {\underset {\color {red}n\times n}{\exp[\mathbf {A^{T}} ]}}={\underset {\color {red}n\times n}{\mathbf {I^{T}} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A^{T}} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}[\mathbf {A^{T}} ]^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {[} {A^{T}}]^{k}}}}
or
exp
[
A
T
]
n
×
n
=
I
n
×
n
+
1
1
!
A
T
n
×
n
+
1
2
!
[
A
T
]
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
[
A
T
]
k
n
×
n
{\displaystyle {\underset {\color {red}n\times n}{\exp[\mathbf {A^{T}} ]}}={\underset {\color {red}n\times n}{\mathbf {I} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A^{T}} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}[\mathbf {A^{T}} ]^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {[} {A^{T}}]^{k}}}}
(Eq.5.1.3 )
Comparing (Eq. 5.1.2 ) and (Eq. 5.1.3 )
We conclude the LHS = RHS, Hence Proved.
R5.2. Exponentiation of a Complex Diagonal Matrix [2] Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
A Diagonal Matrix
D
=
Diag
[
d
1
,
d
2
,
d
3
,
.
.
.
.
,
d
n
]
{\displaystyle \displaystyle \mathbf {D} ={\text{Diag}}[d_{1},d_{2},d_{3},....,d_{n}]}
, where,
D
∈
C
n
×
n
{\displaystyle \mathbf {D} \in \mathbb {C} ^{n\times n}}
.
(Eq.(2)p.20-2b )
Show that
exp
[
D
]
=
Diag
[
e
d
1
,
e
d
2
,
e
d
3
,
.
.
.
.
,
e
d
n
]
{\displaystyle \displaystyle \exp[\mathbf {D} ]={\text{Diag}}[e^{d_{1}},e^{d_{2}},e^{d_{3}},....,e^{d_{n}}]}
, where,
exp
[
D
]
∈
C
n
×
n
{\displaystyle \exp[\mathbf {D} ]\in \mathbb {C} ^{n\times n}}
.
(Eq.(3)p.20-2b )
We know, from Lecture Notes [3] ,
exp
[
A
]
n
×
n
=
I
n
×
n
+
1
1
!
A
n
×
n
+
1
2
!
A
2
n
×
n
+
.
.
.
.
=
∑
k
=
0
∞
1
k
!
A
k
n
×
n
{\displaystyle {\underset {\color {red}n\times n}{\exp[\mathbf {A} ]}}={\underset {\color {red}n\times n}{\mathbf {I} }}+{\underset {\color {red}n\times n}{{\frac {1}{1!}}\mathbf {A} }}+{\underset {\color {red}n\times n}{{\frac {1}{2!}}\mathbf {A} ^{2}}}+....={\underset {\color {red}n\times n}{\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {A} ^{k}}}}
(Eq.(2)p.15-3 )
Let us consider a Simple yet Generic 4x4 Complex Diagonal Matrix
D
{\displaystyle \mathbf {D} }
D
=
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
{\displaystyle \mathbf {D} ={\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}}
(Eq.5.2.2 )
where
i
=
−
1
{\displaystyle i={\sqrt {-1}}}
.
Applying (Eq.(2)p.15-3 ) to (Eq.5.2.2 ) and expanding,
exp
[
D
]
=
[
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
1
]
⏟
Matrix 1
+
1
1
!
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
⏟
Matrix 2
+
1
2
!
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
∗
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
⏟
Matrix 3
+
…
+
1
k
!
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
∗
…
∗
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
⏟
multiply k times - Matrix k
{\displaystyle \exp[\mathbf {D} ]=\underbrace {\begin{bmatrix}1&0&0&0\\0&1&0&0\\0&0&1&0\\0&0&0&1\end{bmatrix}} _{\color {blue}{\text{Matrix 1}}}+{\frac {1}{1!}}\underbrace {\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}} _{\color {blue}{\text{Matrix 2}}}+{\frac {1}{2!}}\underbrace {{\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}*{\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}} _{\color {blue}{\text{Matrix 3}}}+\ldots +{\frac {1}{k!}}\underbrace {{\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}*\ldots *{\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}} _{\color {blue}{\text{multiply k times - Matrix k}}}}
(Eq.5.2.3 )
Simplifying Term 2 and other higher power terms (upto Term k) in the following way,
Matrix 3
=
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
∗
[
a
i
0
0
0
0
b
i
0
0
0
0
c
i
0
0
0
0
d
i
]
{\displaystyle {\text{Matrix 3}}={\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}*{\begin{bmatrix}ai&0&0&0\\0&bi&0&0\\0&0&ci&0\\0&0&0&di\end{bmatrix}}}
Matrix 3
=
[
(
a
i
)
2
0
0
0
0
(
b
i
)
2
0
0
0
0
(
c
i
)
2
0
0
0
0
(
d
i
)
2
]
{\displaystyle {\text{Matrix 3}}={\begin{bmatrix}(ai)^{2}&0&0&0\\0&(bi)^{2}&0&0\\0&0&(ci)^{2}&0\\0&0&0&(di)^{2}\end{bmatrix}}}
(Eq.5.2.4 )
Similarly,
Matrix k
=
[
(
a
i
)
k
0
0
0
0
(
b
i
)
k
0
0
0
0
(
c
i
)
k
0
0
0
0
(
d
i
)
k
]
{\displaystyle {\text{Matrix k}}={\begin{bmatrix}(ai)^{k}&0&0&0\\0&(bi)^{k}&0&0\\0&0&(ci)^{k}&0\\0&0&0&(di)^{k}\end{bmatrix}}}
(Eq.5.2.5 )
Using (Eq.5.2.4 ) and (Eq.5.2.5 ) in (Eq.5.2.3 ) and carrying out simple matrix addition, we get,
e
x
p
[
D
]
=
[
(
1
+
1
1
!
(
a
i
)
+
1
2
!
(
a
i
)
2
+
…
+
1
k
!
(
a
i
)
k
0
0
0
0
(
1
+
1
1
!
(
b
i
)
+
1
2
!
(
b
i
)
2
+
…
+
1
k
!
(
b
i
)
k
)
0
0
0
0
(
1
+
1
1
!
(
c
i
)
+
1
2
!
(
c
i
)
2
+
…
+
1
k
!
(
c
i
)
k
)
0
0
0
0
(
1
+
1
1
!
(
d
i
)
+
1
2
!
(
d
i
)
2
+
…
+
1
k
!
(
d
i
)
k
)
]
{\displaystyle exp[\mathbf {D} ]={\begin{bmatrix}(1+{\frac {1}{1!}}(ai)+{\frac {1}{2!}}(ai)^{2}+\ldots +{\frac {1}{k!}}(ai)^{k}&0&0&0\\0&(1+{\frac {1}{1!}}(bi)+{\frac {1}{2!}}(bi)^{2}+\ldots +{\frac {1}{k!}}(bi)^{k})&0&0\\0&0&(1+{\frac {1}{1!}}(ci)+{\frac {1}{2!}}(ci)^{2}+\ldots +{\frac {1}{k!}}(ci)^{k})&0\\0&0&0&(1+{\frac {1}{1!}}(di)+{\frac {1}{2!}}(di)^{2}+\ldots +{\frac {1}{k!}}(di)^{k})\end{bmatrix}}}
(Eq.5.2.6 )
But every diagonal term of the matrix is of the form,
exp
(
x
)
=
1
+
1
1
!
x
+
1
2
!
x
2
+
…
+
1
n
!
x
n
{\displaystyle \exp(x)=1+{\frac {1}{1!}}x+{\frac {1}{2!}}x^{2}+\ldots +{\frac {1}{n!}}x^{n}}
(Eq.5.2.7 )
Therefore, (Eq.5.2.6 ) can be rewritten as,
e
x
p
[
D
]
=
[
e
x
p
(
a
i
)
0
0
0
0
e
x
p
(
b
i
)
0
0
0
0
e
x
p
(
c
i
)
0
0
0
0
e
x
p
(
d
i
)
]
=
[
e
a
i
0
0
0
0
e
b
i
0
0
0
0
e
c
i
0
0
0
0
e
d
i
]
{\displaystyle exp[\mathbf {D} ]={\begin{bmatrix}exp(ai)&0&0&0\\0&exp(bi)&0&0\\0&0&exp(ci)&0\\0&0&0&exp(di)\end{bmatrix}}={\begin{bmatrix}e^{ai}&0&0&0\\0&e^{bi}&0&0\\0&0&e^{ci}&0\\0&0&0&e^{di}\end{bmatrix}}}
(Eq.5.2.8 )
(ai)
,
(bi)
,
(ci)
,
(di)
{\displaystyle {\text{(ai)}},{\text{(bi)}},{\text{(ci)}},{\text{(di)}}}
are nothing but the diagonal elements of the original matrix in (Eq.5.2.2 ). Hence,
exp
[
D
]
=
Diag
[
e
d
1
,
e
d
2
,
e
d
3
,
e
d
4
]
{\displaystyle \displaystyle \exp[\mathbf {D} ]={\text{Diag}}[e^{d_{1}},e^{d_{2}},e^{d_{3}},e^{d_{4}}]}
(Eq.5.2.9 )
Similarly it can be easily found for an
n
×
n
n\times n
complex diagonal matrix that
exp
[
D
]
=
Diag
[
e
d
1
,
e
d
2
,
e
d
3
,
.
.
.
.
,
e
d
n
]
{\displaystyle \displaystyle \exp[\mathbf {D} ]={\text{Diag}}[e^{d_{1}},e^{d_{2}},e^{d_{3}},....,e^{d_{n}}]}
(Eq.5.2.10 )
Hence Proved.
R5.3 Show form of Exponentiation of Matrix in terms of Eigenvalues of that matrix Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
A matrix
A
{\displaystyle \mathbf {A} }
can be decomposed as
A
=
Φ
Λ
Φ
−
1
{\displaystyle \displaystyle \mathbf {A} =\mathbf {\Phi } \,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1}}
(Eq.(2) p.20-4 )
where
Λ
{\displaystyle \mathbf {\Lambda } }
is the diagonal matrix of eigenvalues of matrix
A
{\displaystyle \mathbf {A} }
Λ
=
Diag
[
λ
1
,
λ
2
,
…
,
d
n
]
∈
C
n
×
n
{\displaystyle \displaystyle \mathbf {\Lambda } ={\text{Diag}}[\lambda _{1},\lambda _{2},\ldots ,d_{n}]\in \mathbb {C} ^{n\times n}}
(Eq.(5) p.20-3 )
and
Φ
{\displaystyle \mathbf {\Phi } }
is the matrix established by n linearly independent eigenvectors
ϕ
i
(
i
=
1
,
2
,
3
,
…
,
n
)
{\displaystyle {\boldsymbol {\phi }}_{i}(i=1,2,3,\ldots ,n)}
of matrix
A
{\displaystyle \mathbf {A} }
, that is,
Φ
=
[
ϕ
1
,
ϕ
2
,
…
,
ϕ
n
]
∈
C
n
×
n
{\displaystyle \displaystyle \mathbf {\Phi } =[{\boldsymbol {\phi }}_{1},{\boldsymbol {\phi }}_{2},\ldots ,{\boldsymbol {\phi }}_{n}]\in \mathbb {C} ^{n\times n}}
(Eq.(2) p.20-3 )
Show that
exp
[
A
]
=
Φ
Diag
[
e
λ
1
,
e
λ
2
,
…
,
e
λ
n
]
Φ
−
1
{\displaystyle \displaystyle \exp[\mathbf {A} ]=\mathbf {\Phi } \,{\text{Diag}}[\,e^{\lambda _{1}},\,e^{\lambda _{2}},\ldots ,\,e^{\lambda _{n}}]\mathbf {\Phi } ^{-1}}
The power series expansion of exponentiation of matrix
A
{\displaystyle \mathbf {A} }
in terms of that matrix has been given as
exp
[
A
]
=
I
+
1
1
!
A
+
1
2
!
A
2
+
1
3
!
A
3
+
⋯
=
∑
k
=
0
∞
1
k
!
A
k
{\displaystyle \displaystyle \exp[\mathbf {A} ]=\mathbf {I} +{\frac {1}{1!}}\mathbf {A} +{\frac {1}{2!}}\mathbf {A} ^{2}+{\frac {1}{3!}}\mathbf {A} ^{3}+\cdots =\sum _{k=0}^{\infty }{\frac {1}{k!}}\mathbf {A} ^{k}}
(Eq.(2) p.15-3 )
Since matrix
A
{\displaystyle \mathbf {A} }
can be decomposed as,
A
=
Φ
Λ
Φ
−
1
{\displaystyle \displaystyle \mathbf {A} =\mathbf {\Phi } \,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1}}
(Eq.(2) p.20-4 )
Expanding the
k
{\displaystyle k}
power of matrix
A
{\displaystyle \mathbf {A} }
yields
A
k
=
(
Φ
Λ
Φ
−
1
)
(
Φ
Λ
Φ
−
1
)
⋯
(
Φ
Λ
Φ
−
1
)
⏟
number of factors
(
Φ
Λ
Φ
−
1
)
is k
{\displaystyle \displaystyle \mathbf {A} ^{k}=\underbrace {(\mathbf {\Phi } \,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1})(\mathbf {\Phi } \,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1})\cdots (\mathbf {\Phi } \,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1})} _{{\text{number of factors}}(\mathbf {\Phi } \,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1}){\text{is k}}}}
(Eq.5.3.1 )
Where the factors
(
Φ
)
{\displaystyle (\mathbf {\Phi } )}
which are neighbors of factors
(
Φ
−
1
)
{\displaystyle (\mathbf {\Phi } ^{-1})}
can be all cancelled in pairs, that is,
A
k
=
Φ
Λ
Φ
−
1
Φ
I
Λ
Φ
−
1
Φ
I
Λ
⋯
Λ
Φ
−
1
Φ
I
Λ
Φ
−
1
{\displaystyle \displaystyle \mathbf {A} ^{k}=\mathbf {\Phi } \,\mathbf {\Lambda } \,{\cancelto {\mathbf {I} }{\mathbf {\Phi } ^{-1}\mathbf {\Phi } \,}}\mathbf {\Lambda } \,{\cancelto {\mathbf {I} }{\mathbf {\Phi } ^{-1}\mathbf {\Phi } \,}}\mathbf {\Lambda } \,\cdots \mathbf {\Lambda } \,{\cancelto {\mathbf {I} }{\mathbf {\Phi } ^{-1}\mathbf {\Phi } \,}}\,\mathbf {\Lambda } \,\mathbf {\Phi } ^{-1}}
(Eq.5.3.2 )
⇒
A
k
=
Φ
Λ
Λ
Λ
⋯
Λ
Λ
⏟
k
Λ
s
Φ
−
1
{\displaystyle \displaystyle \Rightarrow \mathbf {A} ^{k}=\mathbf {\Phi } \,\underbrace {\mathbf {\Lambda } \,\mathbf {\Lambda } \,\mathbf {\Lambda } \,\cdots \mathbf {\Lambda } \,\mathbf {\Lambda } \,} _{{\text{k}}\mathbf {\Lambda } {\text{s}}}\mathbf {\Phi } ^{-1}}
(Eq.5.3.3 )
⇒
A
k
=
Φ
Λ
k
Φ
−
1
{\displaystyle \displaystyle \Rightarrow \mathbf {A} ^{k}=\mathbf {\Phi } \,\mathbf {\Lambda } ^{k}\mathbf {\Phi } ^{-1}}
(Eq.5.3.4 )
Thus, the equation (Eq. 5.3.1 ) can be expressed as
exp
[
A
]
=
∑
k
=
0
∞
1
k
!
[
A
k
]
=
∑
k
=
0
∞
1
k
!
[
Φ
Λ
k
Φ
−
1
]
{\displaystyle \displaystyle \exp[\mathbf {A} ]=\sum _{k=0}^{\infty }{\frac {1}{k!}}[\mathbf {A} ^{k}]=\sum _{k=0}^{\infty }{\frac {1}{k!}}[\mathbf {\Phi } \,\mathbf {\Lambda } ^{k}\mathbf {\Phi } ^{-1}]}
(Eq.5.3.5 )
⇒
exp
[
A
]
=
Φ
∑
k
=
0
∞
1
k
!
[
Λ
k
]
Φ
−
1
{\displaystyle \displaystyle \Rightarrow \exp[\mathbf {A} ]=\mathbf {\Phi } \,\sum _{k=0}^{\infty }{\frac {1}{k!}}[\mathbf {\Lambda } ^{k}]\mathbf {\Phi } ^{-1}}
(Eq.5.3.6 )
According to the equation (Eq.(2) p.15-3 ), now we have,
exp
[
A
]
=
Φ
exp
[
Λ
]
Φ
−
1
{\displaystyle \displaystyle \exp[\mathbf {A} ]=\mathbf {\Phi } \,\exp[\mathbf {\Lambda } ]\mathbf {\Phi } ^{-1}}
(Eq.5.3.7 )
Referring to the conclusion obtained in R5.2, which is
exp
[
D
]
=
Diag
[
e
d
1
,
e
d
2
,
…
,
e
d
n
]
{\displaystyle \displaystyle \exp[\mathbf {D} ]={\text{Diag}}[\,e^{d_{1}},\,e^{d_{2}},\ldots ,\,e^{d_{n}}]}
(Eq.(3) p.20-2b )
Replacing the matrix
D
{\displaystyle \mathbf {D} }
with
Λ
{\displaystyle \mathbf {\Lambda } }
, the elements
d
i
{\displaystyle d_{i}}
with
λ
i
{\displaystyle \lambda _{i}}
,where
i
=
1
,
2
,
3
,
…
,
n
{\displaystyle i=1,2,3,\ldots ,n}
and then substituting into (Eq. 5.3.7 ) yields
exp
[
A
]
=
Φ
Diag
[
e
λ
1
,
e
λ
2
,
…
,
e
λ
n
]
Φ
−
1
{\displaystyle \displaystyle \exp[\mathbf {A} ]=\mathbf {\Phi } \,{\text{Diag}}[\,e^{\lambda _{1}},\,e^{\lambda _{2}},\ldots ,\,e^{\lambda _{n}}]\mathbf {\Phi } ^{-1}}
(Eq.5.3.8 )
R5.4 Show Decomposed Form of Matrix and its Exponentiation Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
Exponentiation of a matrix
A
{\displaystyle \mathbf {A} }
can be decomposed as
exp
[
A
]
=
Φ
Diag
[
e
λ
1
,
e
λ
2
,
…
,
e
λ
n
]
Φ
−
1
{\displaystyle \displaystyle \exp[\mathbf {A} ]=\mathbf {\Phi } \,{\text{Diag}}[\,e^{\lambda _{1}},\,e^{\lambda _{2}},\ldots ,\,e^{\lambda _{n}}]\mathbf {\Phi } ^{-1}}
(Eq.(3) p.20-4 )
The matrix
B
{\displaystyle \mathbf {B} }
is defined in lecture note
B
=
[
0
−
1
1
0
]
{\displaystyle \displaystyle \mathbf {B} ={\begin{bmatrix}0&-1\\1&0\end{bmatrix}}}
(Eq.(1) p.20-2 )
Show
B
t
=
[
1
1
−
i
i
]
[
i
t
0
0
−
i
t
]
[
i
−
1
i
1
]
1
2
i
{\displaystyle \displaystyle \mathbf {B} t={\begin{bmatrix}1&1\\-i&i\end{bmatrix}}{\begin{bmatrix}{it}&0\\0&{-it}\end{bmatrix}}{\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {1}{2i}}}
(Eq.(1) p.20-5 )
and
exp
[
B
t
]
=
[
cos
t
−
sin
t
sin
t
cos
t
]
≠
[
1
e
−
t
e
t
1
]
{\displaystyle \displaystyle \displaystyle \exp[\mathbf {B} t]={\begin{bmatrix}\cos t&-\sin t\\\sin t&\cos t\end{bmatrix}}\neq {\begin{bmatrix}1&e^{-t}\\e^{t}&1\end{bmatrix}}}
(Eq.(2) p.20-5 )
To show the equation (Eq.(1) p.20-5 ), we should first find eigenvalues
λ
{\displaystyle \lambda }
of matrix
B
{\displaystyle \mathbf {B} }
using the matrix equation as follow, indroduce
I
{\displaystyle \mathbf {I} }
to represent Identity matrix.
f
(
x
)
=
|
λ
I
−
B
|
=
0
{\displaystyle \displaystyle f(x)=\left|\lambda \mathbf {I} -\mathbf {B} \right|=0}
⇒
f
(
x
)
=
|
[
λ
0
0
λ
]
−
[
0
−
1
1
0
]
|
=
|
[
λ
1
−
1
λ
]
|
=
0
{\displaystyle \displaystyle \Rightarrow f(x)=\left|{\begin{bmatrix}\lambda &0\\0&\lambda \end{bmatrix}}-{\begin{bmatrix}0&-1\\1&0\end{bmatrix}}\right|=\left|{\begin{bmatrix}\lambda &1\\-1&\lambda \end{bmatrix}}\right|=0}
⇒
f
(
x
)
=
λ
2
+
1
=
0
{\displaystyle \displaystyle \Rightarrow f(x)=\lambda ^{2}+1=0}
⇒
λ
1
=
i
,
λ
2
=
−
i
{\displaystyle \displaystyle \Rightarrow \lambda _{1}=i,\lambda _{2}=-i}
Since the two eigenvalues of matrix
B
{\displaystyle \mathbf {B} }
are both obtained, now solve for the corresponding two eigenvectors.
(
λ
1
I
−
B
)
X
=
O
{\displaystyle \displaystyle (\lambda _{1}\mathbf {I} -\mathbf {B} )\mathbf {X} =\mathbf {O} }
⇒
(
[
λ
1
0
0
λ
1
]
−
[
0
−
1
1
0
]
)
[
x
1
x
2
]
=
[
0
0
]
{\displaystyle \displaystyle \Rightarrow ({\begin{bmatrix}\lambda _{1}&0\\0&\lambda _{1}\end{bmatrix}}-{\begin{bmatrix}0&-1\\1&0\end{bmatrix}}){\begin{bmatrix}x_{1}\\x_{2}\end{bmatrix}}={\begin{bmatrix}0\\0\end{bmatrix}}}
⇒
[
λ
1
1
−
1
λ
1
]
[
x
1
x
2
]
=
[
0
0
]
{\displaystyle \displaystyle \Rightarrow {\begin{bmatrix}\lambda _{1}&1\\-1&\lambda _{1}\end{bmatrix}}{\begin{bmatrix}x_{1}\\x_{2}\end{bmatrix}}={\begin{bmatrix}0\\0\end{bmatrix}}}
(Eq.5.4.1 )
Thus, for the first value of
λ
{\displaystyle \lambda }
, we have
{
λ
1
x
1
+
x
2
=
0
−
x
1
+
λ
1
x
2
=
0
{\displaystyle \displaystyle \left\{{\begin{matrix}\lambda _{1}x_{1}+x_{2}=0\\-x_{1}+\lambda _{1}x_{2}=0\end{matrix}}\right.}
(Eq.5.4.2 )
Substituting
λ
1
=
i
{\displaystyle \lambda _{1}=i}
into the equations above and solving yields, for the eigenvalue
λ
1
=
i
{\displaystyle \lambda _{1}=i}
, that
{
x
1
=
1
x
2
=
−
i
{\displaystyle \displaystyle \left\{{\begin{matrix}x_{1}=1\\x_{2}=-i\end{matrix}}\right.}
(Eq.5.4.3 )
Similarly, we have the equation which can be used for solving eigenvector corresponding to
λ
2
=
−
i
{\displaystyle \lambda _{2}=-i}
,
{
λ
2
x
1
+
x
2
=
0
−
x
1
+
λ
2
x
2
=
0
{\displaystyle \displaystyle \left\{{\begin{matrix}\lambda _{2}x_{1}+x_{2}=0\\-x_{1}+\lambda _{2}x_{2}=0\end{matrix}}\right.}
(Eq.5.4.4 )
Substituting
λ
2
=
−
i
{\displaystyle \lambda _{2}=-i}
into the equations above and solving yields, for the eigenvalue
λ
2
=
−
i
{\displaystyle \lambda _{2}=-i}
, that
{
x
1
=
1
x
2
=
i
{\displaystyle \displaystyle \left\{{\begin{matrix}x_{1}=1\\x_{2}=i\end{matrix}}\right.}
(Eq.5.4.5 )
Now we have obtained two eigenvectors
ϕ
1
{\displaystyle {\boldsymbol {\phi }}_{1}}
and
ϕ
2
{\displaystyle {\boldsymbol {\phi }}_{2}}
of matrix
B
{\displaystyle \mathbf {B} }
, where
ϕ
1
=
[
1
−
i
]
{\displaystyle \displaystyle {\boldsymbol {\phi }}_{1}={\begin{bmatrix}1\\-i\end{bmatrix}}}
,
ϕ
2
=
[
1
i
]
{\displaystyle \displaystyle {\boldsymbol {\phi }}_{2}={\begin{bmatrix}1\\i\end{bmatrix}}}
(Eq.5.4.6 )
Thus we have
Φ
=
[
ϕ
1
,
ϕ
2
]
=
[
1
1
−
i
i
]
{\displaystyle \displaystyle \mathbf {\Phi } =[{\boldsymbol {\phi }}_{1},{\boldsymbol {\phi }}_{2}]={\begin{bmatrix}1&1\\-i&i\end{bmatrix}}}
(Eq.5.4.7 )
Then, calculating the inverse matrix of matrix
Φ
{\displaystyle \mathbf {\Phi } }
yields
Φ
−
1
=
[
i
−
1
i
1
]
1
2
i
{\displaystyle \displaystyle \mathbf {\Phi } ^{-1}={\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {1}{2i}}}
(Eq.5.4.8 )
Therefore we reach the conclusion that,
B
=
[
1
1
−
i
i
]
[
i
0
0
−
i
]
[
i
−
1
i
1
]
1
2
i
{\displaystyle \displaystyle \mathbf {B} ={\begin{bmatrix}1&1\\-i&i\end{bmatrix}}{\begin{bmatrix}{i}&0\\0&{-i}\end{bmatrix}}{\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {1}{2i}}}
⇒
B
t
=
[
1
1
−
i
i
]
[
i
t
0
0
−
i
t
]
[
i
−
1
i
1
]
1
2
i
{\displaystyle \displaystyle \Rightarrow \mathbf {B} t={\begin{bmatrix}1&1\\-i&i\end{bmatrix}}{\begin{bmatrix}{it}&0\\0&{-it}\end{bmatrix}}{\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {1}{2i}}}
(Eq.(1) p.20-5 )
According to the conclusion we have reached in R5.3, we have,
exp
[
B
]
=
Φ
Diag
[
e
λ
1
,
e
λ
2
]
Φ
−
1
{\displaystyle \displaystyle \exp[\mathbf {B} ]=\mathbf {\Phi } \,{\text{Diag}}[\,e^{\lambda _{1}},\,e^{\lambda _{2}}]\mathbf {\Phi } ^{-1}}
⇒
exp
[
B
t
]
=
[
1
1
−
i
i
]
[
e
i
0
0
e
−
i
]
[
i
−
1
i
1
]
t
2
i
{\displaystyle \displaystyle \Rightarrow \exp[\mathbf {B} t]={\begin{bmatrix}1&1\\-i&i\end{bmatrix}}{\begin{bmatrix}e^{i}&0\\0&e^{-i}\end{bmatrix}}{\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {t}{2i}}}
⇒
exp
[
B
t
]
=
[
1
1
−
i
i
]
[
e
i
t
0
0
e
−
i
t
]
[
i
−
1
i
1
]
1
2
i
{\displaystyle \displaystyle \Rightarrow \exp[\mathbf {B} t]={\begin{bmatrix}1&1\\-i&i\end{bmatrix}}{\begin{bmatrix}e^{it}&0\\0&e^{-it}\end{bmatrix}}{\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {1}{2i}}}
Doing the multiplication of matrices at the right side of equation above yields
exp
[
B
t
]
=
[
e
i
t
e
−
i
t
−
i
e
i
t
i
e
−
i
t
]
[
i
−
1
i
1
]
1
2
i
{\displaystyle \displaystyle \exp[\mathbf {B} t]={\begin{bmatrix}e^{it}&e^{-it}\\-ie^{it}&ie^{-it}\end{bmatrix}}{\begin{bmatrix}i&-1\\i&1\end{bmatrix}}{\frac {1}{2i}}}
⇒
exp
[
B
t
]
=
[
i
e
i
t
+
i
e
−
i
t
e
−
i
t
−
e
i
t
e
i
t
−
e
−
i
t
−
i
e
−
i
t
+
i
e
i
t
]
1
2
i
{\displaystyle \displaystyle \Rightarrow \exp[\mathbf {B} t]={\begin{bmatrix}ie^{it}+ie^{-it}&e^{-it}-e^{it}\\e^{it}-e^{-it}&-ie^{-it}+ie^{it}\end{bmatrix}}{\frac {1}{2i}}}
⇒
exp
[
B
t
]
=
[
1
2
(
e
i
t
+
e
−
i
t
)
1
2
i
(
e
−
i
t
−
e
i
t
)
1
2
i
(
e
i
t
−
e
−
i
t
)
1
2
(
e
i
t
+
e
−
i
t
)
]
{\displaystyle \displaystyle \Rightarrow \exp[\mathbf {B} t]={\begin{bmatrix}{\frac {1}{2}}(e^{it}+e^{-it})&{\frac {1}{2i}}(e^{-it}-e^{it})\\{\frac {1}{2i}}(e^{it}-e^{-it})&{\frac {1}{2}}(e^{it}+e^{-it})\end{bmatrix}}}
(Eq.5.4.9 )
Consider Euler’s Formula,[4]
e
i
t
=
cos
t
+
i
sin
t
{\displaystyle \displaystyle e^{it}=\cos t+i\sin t}
(Eq.5.4.10 )
Replacing
i
{\displaystyle \displaystyle i}
with
−
i
{\displaystyle \displaystyle -i}
yields
e
−
i
t
=
cos
t
−
i
sin
t
{\displaystyle \displaystyle e^{-it}=\cos t-i\sin t}
(Eq.5.4.11 )
Solve (Eq.5.4.10 ) together with (Eq.5.4.11 ), we have
{
e
i
t
=
cos
t
+
i
sin
t
e
−
i
t
=
cos
t
−
i
sin
t
{\displaystyle \displaystyle \left\{{\begin{matrix}e^{it}=\cos t+i\sin t\\e^{-it}=\cos t-i\sin t\end{matrix}}\right.}
⇒
{
cos
t
=
1
2
(
e
i
t
+
e
−
i
t
)
sin
t
=
1
2
i
(
e
i
t
−
e
−
i
t
)
{\displaystyle \displaystyle \Rightarrow \left\{{\begin{matrix}\cos t={\frac {1}{2}}(e^{it}+e^{-it})\\\sin t={\frac {1}{2i}}(e^{it}-e^{-it})\end{matrix}}\right.}
(Eq.5.4.12 )
Substituting (Eq.5.4.12 ) into (Eq.5.4.9 ) yields
exp
[
B
t
]
=
[
cos
t
−
sin
t
sin
t
cos
t
]
{\displaystyle \displaystyle \exp[\mathbf {B} t]={\begin{bmatrix}\cos t&-\sin t\\\sin t&\cos t\end{bmatrix}}}
(Eq.(2) p.20-5 )
Obviously,
[
cos
t
−
sin
t
sin
t
cos
t
]
≠
[
1
e
−
t
e
t
1
]
{\displaystyle \displaystyle {\begin{bmatrix}\cos t&-\sin t\\\sin t&\cos t\end{bmatrix}}\neq {\begin{bmatrix}1&e^{-t}\\e^{t}&1\end{bmatrix}}}
R*5.5 Generating a class of exact L2-ODE-VC [5] Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
A L2-ODE-VC [6] :
G
=
R
(
x
)
y
+
Q
(
x
)
y
′
+
P
(
x
)
y
″
{\displaystyle \displaystyle G=R(x)y+Q(x)y^{\prime }+P(x)y^{''}}
(Eq. 5.5.1 )
The first intregal
ϕ
(
x
,
y
,
p
)
{\displaystyle \phi (x,y,p)}
can also be expressed as:
ϕ
x
(
x
,
y
,
p
)
+
ϕ
y
(
x
,
y
,
p
)
y
′
=
R
(
x
)
y
+
Q
(
x
)
y
′
{\displaystyle \displaystyle \phi _{x}(x,y,p)+\phi _{y}(x,y,p)y^{\prime }=R(x)y+Q(x)y^{\prime }}
(Eq. 5.5.2 )
Show that(Eq. 5.5.1 ) and (Eq. 5.5.2 ) lead to a general class of exact L2-ODE-VC of the form:
ϕ
(
x
,
y
,
p
)
=
P
(
x
)
p
+
T
(
x
)
y
+
k
{\displaystyle \displaystyle \phi (x,y,p)=P(x)p+T(x)y+k}
(Eq. 5.5.3 )
Nomenclature Edit
p
:=
y
′
:=
d
y
d
x
{\displaystyle \displaystyle p:=y':={\frac {dy}{dx}}}
Derivation of Eq. 5.5.3 Edit
The first exactness condition for L2-ODE-VC: [8]
G
=
d
ϕ
(
x
,
y
,
p
)
d
x
=
ϕ
x
+
ϕ
y
p
+
ϕ
p
y
″
{\displaystyle \displaystyle G={\frac {d\phi (x,y,p)}{dx}}=\phi _{x}+\phi _{y}p+\phi _{p}y^{''}}
(Eq. 5.5.4 )
From (Eq. 5.5.1 ) and (Eq. 5.5.4 ), we can infer that
ϕ
p
=
P
(
x
)
{\displaystyle \displaystyle \phi _{p}=P(x)}
(Eq. 5.5.5 )
Integrating (Eq. 5.5.5 ), w.r.t p, we obtain:
ϕ
=
P
(
x
)
p
+
k
(
x
,
y
)
{\displaystyle \displaystyle \phi =P(x)p+k(x,y)}
(Eq. 5.5.6 )
Partial derivatives of
ϕ
{\displaystyle \phi }
w.r.t to x and y can be written as:
ϕ
x
=
P
′
(
x
)
p
+
∂
k
(
x
,
y
)
∂
x
{\displaystyle \displaystyle \phi _{x}=P^{\prime }(x)p+{\frac {\partial k(x,y)}{\partial x}}}
(Eq. 5.5.7 )
ϕ
y
=
∂
k
(
x
,
y
)
∂
y
{\displaystyle \displaystyle \phi _{y}={\frac {\partial k(x,y)}{\partial y}}}
(Eq. 5.5.8 )
Substituting the partial derivatives of
ϕ
\phi
w.r.t x,y and p [(Eq. 5.5.7 ), (Eq. 5.5.8 ), (Eq. 5.5.6 )] into (Eq. 5.5.4 ), we obtain:
G
=
P
′
(
x
)
p
+
∂
k
(
x
,
y
)
∂
y
p
+
∂
k
(
x
,
y
)
∂
x
+
P
(
x
)
y
″
{\displaystyle \displaystyle G=P^{\prime }(x)p+{\frac {\partial k(x,y)}{\partial y}}p+{\frac {\partial k(x,y)}{\partial x}}+P(x)y^{''}}
(Eq. 5.5.8 )
Comparing (Eq. 5.5.8 ) with (Eq. 5.5.1 ), we can write:
G
=
P
(
x
)
y
″
+
(
∂
k
(
x
,
y
)
∂
y
+
P
′
(
x
)
)
⏟
Q
(
x
)
p
+
∂
k
(
x
,
y
)
∂
x
⏟
R
(
x
)
y
{\displaystyle \displaystyle G=P(x)y^{''}+\underbrace {\left({\frac {\partial k(x,y)}{\partial y}}+P^{\prime }(x)\right)} _{Q(x)}p+\underbrace {\frac {\partial k(x,y)}{\partial x}} _{R(x)y}}
(Eq. 5.5.9 )
Thus
∂
k
(
x
,
y
)
∂
x
=
R
(
x
)
y
{\displaystyle \displaystyle {\frac {\partial k(x,y)}{\partial x}}=R(x)y}
Integrating w.r.t x,
k
(
x
,
y
)
=
y
∫
x
R
(
s
)
d
s
⏟
T
(
x
)
+
k
1
(
y
)
{\displaystyle \displaystyle k(x,y)=y\underbrace {\int ^{x}R(s)ds} _{T(x)}+k_{1}(y)}
(Eq. 5.5.10 )
Substituting the
k
(
x
,
y
)
{\displaystyle k(x,y)}
obtained in (Eq. 5.5.10 ) back into the expression for
ϕ
{\displaystyle \phi }
obtained in (Eq. 5.5.6 ), we obtain:
ϕ
=
P
(
x
)
p
+
T
(
x
)
y
+
k
1
(
y
)
{\displaystyle \displaystyle \phi =P(x)p+T(x)y+k_{1}(y)}
(Eq. 5.5.11 )
The partial derivative of
ϕ
\phi
(Eq. 5.5.11 ) w.r.t y,
ϕ
y
=
T
(
x
)
+
k
1
′
(
y
)
{\displaystyle \displaystyle \phi _{y}=T(x)+k_{1}^{\prime }(y)}
(Eq. 5.5.12 )
But from (Eq. 5.5.1 ) and (Eq. 5.5.2 ), we see that
ϕ
y
=
Q
(
x
)
{\displaystyle \displaystyle \phi _{y}=Q(x)}
.
So,
Q
(
x
)
=
T
(
x
)
+
k
1
′
(
y
)
{\displaystyle Q(x)=T(x)+k'_{1}(y)}
Since,
Q
(
x
)
{\displaystyle Q(x)}
is only a function of
x
x
, so, we can now say that
T
(
x
)
=
Q
(
x
)
{\displaystyle T(x)=Q(x)}
and
k
1
′
(
y
)
=
0
{\displaystyle {k_{1}}^{\prime }(y)=0}
.
Thus
k
1
(
y
)
{\displaystyle k_{1}(y)}
is a constant.
Hence we obtain the following expression for
ϕ
{\displaystyle \phi }
:
ϕ
(
x
,
y
,
p
)
=
P
(
x
)
p
+
T
(
x
)
y
+
k
{\displaystyle \displaystyle \phi (x,y,p)=P(x)p+T(x)y+k}
(Eq. 5.5.13 )
which represents a general class of Exact L2-ODE-VC.
R*5.6 Solving a L2-ODE-VC [9] Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
G
=
(
c
o
s
x
)
y
″
+
(
x
2
−
s
i
n
x
)
y
′
+
2
x
y
=
0
{\displaystyle \displaystyle G=(cosx)y''+(x^{2}-sinx)y'+2xy=0}
(Eq. 5.6.1 )
1. Show that (Eq. 5.6.1 ) is exact.
2. Find
ϕ
{\displaystyle \displaystyle \phi }
3. Solve for
y
(
x
)
{\displaystyle \displaystyle y(x)}
Nomenclature Edit
p
=:
y
′
=:
d
y
d
x
{\displaystyle \displaystyle p=:y'=:{\frac {dy}{dx}}}
f
i
j
=
d
2
f
∂
i
∂
j
{\displaystyle \displaystyle f_{ij}={\frac {d^{2}f}{\partial i\partial j}}}
Exactness Conditions[10] Edit
The exactness conditions for N2-ODE (Non Linear Second Order Differential Equation) are:
First Exactness condition
For an equation to be exact, they must be of the form
G
(
x
,
y
,
y
′
,
y
″
)
=
g
(
x
,
y
,
p
)
+
f
(
x
,
y
,
p
)
y
″
=
d
ϕ
d
x
{\displaystyle \displaystyle G(x,y,y',y'')=g(x,y,p)+f(x,y,p)y''={\frac {\mathrm {d} \phi }{\mathrm {d} x}}}
g
(
x
,
y
,
p
)
=
ϕ
x
+
ϕ
y
.
y
′
{\displaystyle \displaystyle g(x,y,p)=\phi _{x}+\phi _{y}.y'}
(Eq. 5.6.2 )
f
(
x
,
y
,
p
)
=
ϕ
p
{\displaystyle \displaystyle f(x,y,p)=\phi _{p}}
(Eq. 5.6.3 )
Second Exactness Condition
f
x
x
+
2
p
f
x
y
+
p
2
f
y
y
=
g
x
p
+
p
g
y
p
−
g
y
{\displaystyle f_{xx}+2pf_{xy}+p^{2}f_{yy}=g_{xp}+pg_{yp}-g_{y}}
(Eq. 5.6.4 )
f
x
p
+
p
f
y
p
+
2
f
y
=
g
p
p
{\displaystyle f_{xp}+pf_{yp}+2f_{y}=g_{pp}}
(Eq. 5.6.5 )
We have
G
=
(
c
o
s
x
)
y
″
+
(
x
2
−
s
i
n
x
)
y
′
+
2
x
y
=
0
{\displaystyle \displaystyle G=(cosx)y''+(x^{2}-sinx)y'+2xy=0}
Where we can identify
g
(
x
,
y
,
p
)
=
(
x
2
−
s
i
n
x
)
y
′
+
2
x
y
{\displaystyle \displaystyle g(x,y,p)=(x^{2}-sinx)y'+2xy}
and
f
(
x
,
y
,
p
)
=
c
o
s
x
{\displaystyle \displaystyle f(x,y,p)=cosx}
Thus the equation satisfies the first exactness condition.
For the second exactness condition, we first calculate the various partial derivatives of f and g.
g
p
=
x
2
−
s
i
n
x
{\displaystyle \displaystyle g_{p}=x^{2}-sinx}
g
p
p
=
0
{\displaystyle \displaystyle g_{pp}=0}
g
x
p
=
2
x
−
c
o
s
x
{\displaystyle \displaystyle g_{xp}=2x-cosx}
g
y
p
=
0
{\displaystyle \displaystyle g_{yp}=0}
g
y
=
2
x
{\displaystyle \displaystyle g_{y}=2x}
f
p
=
0
{\displaystyle \displaystyle f_{p}=0}
f
x
p
=
f
y
p
=
0
{\displaystyle \displaystyle f_{xp}=f_{yp}=0}
f
y
=
f
y
y
=
f
x
y
=
0
{\displaystyle \displaystyle f_{y}=f_{yy}=f_{xy}=0}
f
x
=
−
s
i
n
x
{\displaystyle \displaystyle f_{x}=-sinx}
f
x
x
=
−
c
o
s
x
{\displaystyle \displaystyle f_{xx}=-cosx}
Substituting the values in (Eq. 5.6.4 ) we get
L
.
H
.
S
=
−
c
o
s
x
+
0
+
0
=
−
c
o
s
x
{\displaystyle \displaystyle L.H.S=-cosx+0+0=-cosx}
R
.
H
.
S
=
2
x
−
c
o
s
x
+
0
−
2
x
=
−
c
o
s
x
{\displaystyle \displaystyle R.H.S=2x-cosx+0-2x=-cosx}
Therefore the first equation satisfies.
Substituting the values in (Eq. 5.6.5 ) we get
L
.
H
.
S
.
=
0
+
0
+
0
=
0
{\displaystyle \displaystyle L.H.S.=0+0+0=0}
R
.
H
.
S
=
0
{\displaystyle \displaystyle R.H.S=0}
Therefore the second equation satisfies as well.
Thus the second exactness condition is satisfied and the given differential equation is exact.
Now, we have
f
(
x
,
y
,
p
)
=
ϕ
p
{\displaystyle \displaystyle f(x,y,p)=\phi _{p}}
Integrating w.r.t. p, we get
ϕ
=
∫
c
o
s
x
.
d
p
+
h
(
x
,
y
)
{\displaystyle \displaystyle \phi =\int cosx.dp+h(x,y)}
where h(x,y) is a function of integration as we integrated only partially w.r.t. p.
ϕ
=
p
.
c
o
s
x
+
h
(
x
,
y
)
{\displaystyle \displaystyle \phi =p.cosx+h(x,y)}
(Eq. 5.6.6 )
Partially differentiating (Eq. 5.6.6 ) w.r.t x
ϕ
x
=
−
p
s
i
n
x
.
+
h
x
{\displaystyle \displaystyle \phi _{x}=-psinx.+h_{x}}
Partially differentiating (Eq. 5.6.6 ) w.r.t y
ϕ
y
=
h
y
{\displaystyle \displaystyle \phi _{y}=h_{y}}
From equation (Eq. 5.6.3 ), we have
g
(
x
,
y
,
p
)
=
ϕ
x
+
ϕ
y
.
y
′
{\displaystyle \displaystyle g(x,y,p)=\phi _{x}+\phi _{y}.y'}
=
−
p
.
s
i
n
x
+
h
x
+
h
y
.
p
{\displaystyle \displaystyle =-p.sinx+h_{x}+h_{y}.p}
=
p
(
h
y
−
s
i
n
x
)
+
h
x
{\displaystyle \displaystyle =p(h_{y}-sinx)+h_{x}}
We have established that
g
(
x
,
y
,
p
)
=
(
x
2
−
s
i
n
x
)
y
′
+
2
x
y
{\displaystyle \displaystyle g(x,y,p)=(x^{2}-sinx)y'+2xy}
Comparing the two equations, we get,
h
x
=
2
x
y
{\displaystyle \displaystyle h_{x}=2xy}
On integrating,
h
=
x
2
y
+
k
1
(
x
)
{\displaystyle \displaystyle h=x^{2}y+k_{1}(x)}
Thus,
h
x
=
2
x
y
+
k
1
′
=
2
x
y
,
∴
k
1
=
c
o
n
s
t
{\displaystyle \displaystyle h_{x}=2xy+k_{1}'=2xy,\therefore k_{1}=const}
Thus we have
ϕ
=
c
o
s
x
.
y
′
+
x
2
y
+
k
1
=
k
2
{\displaystyle \displaystyle \phi =cosx.y'+x^{2}y+k_{1}=k_{2}}
ϕ
=
c
o
s
x
.
y
′
+
x
2
y
=
k
{\displaystyle \displaystyle \phi =cosx.y'+x^{2}y=k}
y
′
+
x
2
y
c
o
s
x
=
k
c
o
s
x
{\displaystyle \displaystyle y'+{\frac {x^{2}y}{cosx}}={\frac {k}{cosx}}}
This N1-ODE can be solved using the Integrating Factor Method that we very well know.
h
(
x
)
=
e
∫
x
2
c
o
s
x
.
d
x
{\displaystyle \displaystyle h(x)=e^{\int {\frac {x^{2}}{cosx}}.dx}}
y
=
1
h
(
x
)
.
∫
h
(
x
)
.
k
c
o
s
x
.
d
x
+
c
{\displaystyle \displaystyle y={\frac {1}{h(x)}}.\int h(x).{\frac {k}{cosx}}.dx+c}
y
=
1
e
∫
x
2
c
o
s
x
.
d
x
∫
[
e
∫
x
2
c
o
s
x
.
d
x
k
c
o
s
x
d
x
]
+
c
{\displaystyle \displaystyle y={\frac {1}{e^{\int {\frac {x^{2}}{cosx}}.dx}}}\int \left[e^{\int {\frac {x^{2}}{cosx}}.dx}{\frac {k}{cosx}}dx\right]+c}
R*5.7 Show equivalence to symmetry of second partial derivatives of first integral [11] Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
g
0
−
d
g
1
d
x
+
d
2
g
2
d
x
2
=
0
{\displaystyle \displaystyle g_{0}-{\frac {dg_{1}}{dx}}+{\frac {d^{2}g_{2}}{dx^{2}}}=0}
(Eq.(1) p.22-3 )
where
g
i
:=
∂
G
∂
y
(
i
)
=
∂
∂
y
(
i
)
(
d
ϕ
d
x
)
,
f
o
r
i
=
0
,
1
,
2
{\displaystyle \displaystyle g_{i}:={\frac {\partial G}{\partial y^{(i)}}}={\frac {\partial }{\partial y^{(i)}}}({\frac {d\phi }{dx}}),{\rm {{for}\ i=0,1,2}}}
(Eq.(3) p.21-7 )
Show equivalence to symmetry of mixed second partial derivatives of first integral, that is
ϕ
x
y
=
ϕ
y
x
,
ϕ
p
y
=
ϕ
y
p
,
ϕ
x
p
=
ϕ
p
x
{\displaystyle \displaystyle \phi _{xy}=\phi _{yx},\phi _{py}=\phi _{yp},\phi _{xp}=\phi _{px}}
where
p
(
x
)
:=
y
′
(
x
)
{\displaystyle \displaystyle p(x):=y'(x)}
g
0
:=
∂
∂
y
(
d
ϕ
d
x
)
{\displaystyle \displaystyle g_{0}:={\frac {\partial }{\partial y}}({\frac {d\phi }{dx}})}
(Eq.(3) p.21-7 )
d
d
x
g
1
=
d
d
x
(
ϕ
x
p
+
ϕ
y
p
p
+
ϕ
y
+
ϕ
p
p
y
″
)
{\displaystyle \displaystyle {\frac {d}{dx}}g_{1}={\frac {d}{dx}}(\phi _{xp}+\phi _{yp}p+\phi _{y}+\phi _{pp}y'')}
(Eq.(2) p.22-4 )
From (Eq.(2)p.22-4 ), we have,
d
d
x
g
1
=
d
d
x
[
ϕ
x
p
+
ϕ
y
p
y
′
+
ϕ
p
p
y
″
]
+
d
ϕ
y
d
x
{\displaystyle \displaystyle {\frac {d}{dx}}g_{1}={\frac {d}{dx}}[\phi _{xp}+\phi _{yp}y'+\phi _{pp}y'']+{\frac {d\phi _{y}}{dx}}}
⇒
d
d
x
g
1
=
d
d
x
[
ϕ
x
p
+
ϕ
y
p
p
+
ϕ
p
p
y
″
]
+
d
d
x
(
∂
ϕ
∂
y
)
{\displaystyle \displaystyle \Rightarrow {\frac {d}{dx}}g_{1}={\frac {d}{dx}}[\phi _{xp}+\phi _{yp}p+\phi _{pp}y'']+{\frac {d}{dx}}({\frac {\partial \phi }{\partial y}})}
(Eq. 5.7.1 )
d
2
d
x
2
g
2
=
d
d
x
[
ϕ
p
x
+
ϕ
p
y
y
′
+
ϕ
p
p
y
″
]
{\displaystyle \displaystyle {\frac {d^{2}}{dx^{2}}}g_{2}={\frac {d}{dx}}[\phi _{px}+\phi _{py}y'+\phi _{pp}y'']}
(Eq.(3) p.22-4 )
Substituting (Eq.(3)p.21-7 ),(Eq. 5.7.1 ) and (Eq.(3)p.22-4 ) into (Eq.(1)p.22-3 ) yields
∂
∂
y
(
d
ϕ
d
x
)
−
d
d
x
[
ϕ
x
p
+
ϕ
y
p
y
′
+
ϕ
p
p
y
″
]
−
d
d
x
(
∂
ϕ
∂
y
)
+
d
d
x
[
ϕ
p
x
+
ϕ
p
y
y
′
+
ϕ
p
p
y
″
]
=
0
{\displaystyle \displaystyle {\frac {\partial }{\partial y}}({\frac {d\phi }{dx}})-{\frac {d}{dx}}[\phi _{xp}+\phi _{yp}y'+\phi _{pp}y'']-{\frac {d}{dx}}({\frac {\partial \phi }{\partial y}})+{\frac {d}{dx}}[\phi _{px}+\phi _{py}y'+\phi _{pp}y'']=0}
⇒
∂
∂
y
(
d
ϕ
d
x
)
−
d
d
x
(
∂
ϕ
∂
y
)
+
d
d
x
[
(
ϕ
p
x
−
ϕ
x
p
)
+
(
ϕ
p
y
−
ϕ
y
p
)
p
]
=
0
{\displaystyle \displaystyle \Rightarrow {\frac {\partial }{\partial y}}({\frac {d\phi }{dx}})-{\frac {d}{dx}}({\frac {\partial \phi }{\partial y}})+{\frac {d}{dx}}[(\phi _{px}-\phi _{xp})+(\phi _{py}-\phi _{yp})p]=0}
(Eq. 5.7.2 )
Because
∂
∂
y
(
d
ϕ
d
x
)
=
∂
∂
y
(
∂
ϕ
∂
x
+
∂
ϕ
∂
y
d
y
d
x
+
∂
ϕ
∂
y
′
d
y
′
d
x
)
=
ϕ
x
y
+
ϕ
y
y
d
y
d
x
+
ϕ
p
y
d
y
′
d
x
{\displaystyle \displaystyle {\frac {\partial }{\partial y}}({\frac {d\phi }{dx}})={\frac {\partial }{\partial y}}({\frac {\partial \phi }{\partial x}}+{\frac {\partial \phi }{\partial y}}{\frac {dy}{dx}}+{\frac {\partial \phi }{\partial y'}}{\frac {dy'}{dx}})=\phi _{xy}+\phi _{yy}{\frac {dy}{dx}}+\phi _{py}{\frac {dy'}{dx}}}
d
d
x
(
∂
ϕ
∂
y
)
=
[
∂
∂
x
(
∂
ϕ
∂
y
)
+
∂
∂
y
(
∂
ϕ
∂
y
)
d
y
d
x
+
∂
∂
y
′
(
∂
ϕ
∂
y
)
d
y
′
d
x
]
=
ϕ
y
x
+
ϕ
y
y
d
y
d
x
+
ϕ
y
p
d
y
′
d
x
{\displaystyle \displaystyle {\frac {d}{dx}}({\frac {\partial \phi }{\partial y}})=[{\frac {\partial }{\partial x}}({\frac {\partial \phi }{\partial y}})+{\frac {\partial }{\partial y}}({\frac {\partial \phi }{\partial y}}){\frac {dy}{dx}}+{\frac {\partial }{\partial y'}}({\frac {\partial \phi }{\partial y}}){\frac {dy'}{dx}}]=\phi _{yx}+\phi _{yy}{\frac {dy}{dx}}+\phi _{yp}{\frac {dy'}{dx}}}
Thus
∂
∂
y
(
d
ϕ
d
x
)
−
d
d
x
(
∂
ϕ
∂
y
)
=
ϕ
x
y
+
ϕ
y
y
d
y
d
x
+
ϕ
p
y
d
y
′
d
x
−
ϕ
y
x
−
ϕ
y
y
d
y
d
x
−
ϕ
y
p
d
y
′
d
x
=
ϕ
x
y
−
ϕ
y
x
+
(
ϕ
p
y
−
ϕ
y
p
)
y
″
{\displaystyle \displaystyle {\frac {\partial }{\partial y}}({\frac {d\phi }{dx}})-{\frac {d}{dx}}({\frac {\partial \phi }{\partial y}})=\phi _{xy}+\phi _{yy}{\frac {dy}{dx}}+\phi _{py}{\frac {dy'}{dx}}-\phi _{yx}-\phi _{yy}{\frac {dy}{dx}}-\phi _{yp}{\frac {dy'}{dx}}=\phi _{xy}-\phi _{yx}+(\phi _{py}-\phi _{yp})y''}
(Eq. 5.7.3 )
Substituting (Eq. 5.7.3 ) into (Eq. 5.7.2 ) yields
(
ϕ
x
y
−
ϕ
y
x
)
+
(
ϕ
p
y
−
ϕ
y
p
)
y
″
+
d
d
x
[
(
ϕ
p
x
−
ϕ
x
p
)
+
(
ϕ
p
y
−
ϕ
y
p
)
p
]
=
0
{\displaystyle \displaystyle (\phi _{xy}-\phi _{yx})+(\phi _{py}-\phi _{yp})y''+{\frac {d}{dx}}[(\phi _{px}-\phi _{xp})+(\phi _{py}-\phi _{yp})p]=0}
(Eq. 5.7.4 )
Because
d
d
x
(
ϕ
p
y
−
ϕ
y
p
)
y
′
=
(
ϕ
p
y
y
″
+
d
ϕ
p
y
d
x
y
′
)
−
(
ϕ
y
p
y
″
+
d
ϕ
y
p
d
x
y
′
)
{\displaystyle \displaystyle {\frac {d}{dx}}(\phi _{py}-\phi _{yp})y'=(\phi _{py}y''+{\frac {d\phi _{py}}{dx}}y')-(\phi _{yp}y''+{\frac {d\phi _{yp}}{dx}}y')}
(Eq. 5.7.5 )
Substitute (Eq. 5.7.5 ) into (Eq. 5.7.4 ), we have
g
0
−
d
g
1
d
x
+
d
2
g
2
d
x
2
=
(
ϕ
x
y
−
ϕ
y
x
)
+
2
(
ϕ
p
y
−
ϕ
y
p
)
y
″
+
d
d
x
(
ϕ
p
x
−
ϕ
x
p
)
+
y
′
d
d
x
(
ϕ
p
y
−
ϕ
y
p
)
=
0
{\displaystyle \displaystyle g_{0}-{\frac {dg_{1}}{dx}}+{\frac {d^{2}g_{2}}{dx^{2}}}=(\phi _{xy}-\phi _{yx})+2(\phi _{py}-\phi _{yp})y''+{\frac {d}{dx}}(\phi _{px}-\phi _{xp})+y'{\frac {d}{dx}}(\phi _{py}-\phi _{yp})=0}
⇒
(
ϕ
x
y
−
ϕ
y
x
)
+
(
2
y
″
+
y
′
d
d
x
)
(
ϕ
p
y
−
ϕ
y
p
)
+
d
d
x
(
ϕ
p
x
−
ϕ
x
p
)
=
0
{\displaystyle \displaystyle \Rightarrow (\phi _{xy}-\phi _{yx})+(2y''+y'{\frac {d}{dx}})(\phi _{py}-\phi _{yp})+{\frac {d}{dx}}(\phi _{px}-\phi _{xp})=0}
(Eq. 5.7.6 )
Since
y
″
{\displaystyle y''}
and
y
′
{\displaystyle y'}
can be the second and first derivative of any solution function
y
y
of any second order ODE in terms of which the equation
g
0
−
d
g
1
d
x
+
d
2
g
2
d
x
2
=
0
{\displaystyle g_{0}-{\frac {dg_{1}}{dx}}+{\frac {d^{2}g_{2}}{dx^{2}}}=0}
is hold. That is, the factor
2
y
″
+
y
′
d
d
x
{\displaystyle 2y''+y'{\frac {d}{dx}}}
, which consists of two derivatives of solution function and the derivative operater so that depends partly on the solution functin of ODE, can be arbitrary and thus linearly indepent of the derivative operater
d
d
x
{\displaystyle {\frac {d}{dx}}}
, which is a factor of the third term on left hand side of (Eq. 5.7.6 ).
Similarly, comparing the first and the third terms on left hand side of (Eq. 5.7.6 ) yields that the factor 1 (which can be treated as a unit nature number basis of function space) of the first term and the derivative operater (which is another basis of derivative function space) of the third term are linearly independent of each other.
For the left side of (Eq. 5.7.6 ) being zero under any circumstances, we should have,
(
2
y
″
+
y
′
d
d
x
)
(
ϕ
p
y
−
ϕ
y
p
)
=
0
{\displaystyle \displaystyle (2y''+y'{\frac {d}{dx}})(\phi _{py}-\phi _{yp})=0}
(Eq. 5.7.7 )
while
(
ϕ
x
y
−
ϕ
y
x
)
=
0
{\displaystyle \displaystyle (\phi _{xy}-\phi _{yx})=0}
(Eq. 5.7.8 )
d
d
x
(
ϕ
p
x
−
ϕ
x
p
)
=
0
{\displaystyle \displaystyle {\frac {d}{dx}}(\phi _{px}-\phi _{xp})=0}
(Eq. 5.7.9 )
From (Eq. 5.7.7 ),since the factor
2
y
″
+
y
′
d
d
x
{\displaystyle 2y''+y'{\frac {d}{dx}}}
is arbitrary, we obtain,
ϕ
p
y
−
ϕ
y
p
=
0
{\displaystyle \displaystyle \phi _{py}-\phi _{yp}=0}
(Eq. 5.7.10 )
Thus,
ϕ
p
y
=
ϕ
y
p
{\displaystyle \displaystyle \phi _{py}=\phi _{yp}}
(Eq. 5.7.11 )
From(Eq. 5.7.9 ), consider
(
ϕ
p
x
−
ϕ
x
p
)
{\displaystyle (\phi _{px}-\phi _{xp})}
to be also a function of variables x,y and p, which can be represented as
h
(
x
,
y
,
p
)
{\displaystyle h(x,y,p)}
, thus,
d
d
x
h
(
x
,
y
,
p
)
=
∂
∂
x
h
(
x
,
y
,
p
)
+
∂
∂
y
h
(
x
,
y
,
p
)
p
+
∂
∂
p
h
(
x
,
y
,
p
)
p
′
=
0
{\displaystyle \displaystyle {\frac {d}{dx}}h(x,y,p)={\frac {\partial }{\partial x}}h(x,y,p)+{\frac {\partial }{\partial y}}h(x,y,p)p+{\frac {\partial }{\partial p}}h(x,y,p)p'=0}
(Eq. 5.7.12 )
Since the partial derivative opraters
∂
∂
x
,
∂
∂
y
,
∂
∂
p
{\displaystyle {\frac {\partial }{\partial x}},{\frac {\partial }{\partial y}},{\frac {\partial }{\partial p}}}
are linearly independent, we have,
∂
∂
x
h
(
x
,
y
,
p
)
=
0
{\displaystyle \displaystyle {\frac {\partial }{\partial x}}h(x,y,p)=0}
(Eq. 5.7.13 )
∂
∂
y
h
(
x
,
y
,
p
)
p
=
0
{\displaystyle \displaystyle {\frac {\partial }{\partial y}}h(x,y,p)p=0}
(Eq. 5.7.14 )
∂
∂
p
h
(
x
,
y
,
p
)
p
′
=
0
{\displaystyle \displaystyle {\frac {\partial }{\partial p}}h(x,y,p)p'=0}
(Eq. 5.7.15 )
Obviously the only condition by which the three equations above are all satisfied is that the function
h
(
x
,
y
,
p
)
{\displaystyle h(x,y,p)}
is a numerical constant.
Thus, we have
(
ϕ
p
x
−
ϕ
x
p
)
=
C
{\displaystyle \displaystyle (\phi _{px}-\phi _{xp})=C}
(Eq. 5.7.16 )
where
C
{\displaystyle C}
is a constant. To find the value of constant
C
{\displaystyle C}
, try the process as follow.
ϕ
p
x
=
ϕ
x
p
+
C
{\displaystyle \displaystyle \phi _{px}=\phi _{xp}+C}
(Eq. 5.7.17 )
Find integral on both sides of (Eq. 5.7.17 ) in terms of x,
ϕ
p
=
∫
ϕ
x
p
d
x
+
C
x
+
f
(
y
,
p
)
{\displaystyle \displaystyle \phi _{p}=\int \phi _{xp}dx+Cx+f(y,p)}
(Eq. 5.7.18 )
where the term
f
(
y
,
p
)
{\displaystyle f(y,p)}
is an arbitrarily selected function of independent variables y and p. Then find integral on both sides of (Eq. 5.7.18 ) in terms of p,
ϕ
=
∫
∫
ϕ
x
p
d
x
d
p
+
C
x
p
+
∫
f
(
y
,
p
)
d
p
+
g
(
x
,
y
)
{\displaystyle \displaystyle \phi =\int \int \phi _{xp}dxdp+Cxp+\int f(y,p)dp+g(x,y)}
(Eq. 5.7.19 )
where the term
g
(
x
,
y
)
{\displaystyle g(x,y)}
is an arbitrarily selected function of variables x and y.
The first partial derivatives of both sides of (Eq. 5.7.19 ) in terms of x could be
ϕ
x
=
∫
ϕ
x
p
d
p
+
C
p
+
∂
∂
x
[
∫
f
(
y
,
p
)
p
′
d
x
+
g
(
x
,
y
)
]
{\displaystyle \displaystyle \phi _{x}=\int \phi _{xp}dp+Cp+{\frac {\partial }{\partial x}}[\int f(y,p)p'dx+g(x,y)]}
(Eq. 5.7.20 )
⇒
ϕ
x
=
∫
ϕ
x
p
d
p
+
C
p
+
f
(
y
,
p
)
p
′
+
g
′
(
x
,
y
)
{\displaystyle \displaystyle \Rightarrow \phi _{x}=\int \phi _{xp}dp+Cp+f(y,p)p'+g'(x,y)}
(Eq. 5.7.21 )
Then find partial derivative of both sides of (Eq. 5.7.21 ) in terms of p,
ϕ
x
p
=
ϕ
x
p
+
C
+
∂
∂
p
[
f
(
y
,
p
)
p
′
]
{\displaystyle \displaystyle \phi _{xp}=\phi _{xp}+C+{\frac {\partial }{\partial p}}[f(y,p)p']}
(Eq. 5.7.22 )
⇒
C
+
∂
∂
p
[
f
(
y
,
p
)
]
p
′
=
0
{\displaystyle \displaystyle \Rightarrow C+{\frac {\partial }{\partial p}}[f(y,p)]p'=0}
(Eq. 5.7.23 )
⇒
C
p
′
=
−
∂
∂
p
[
f
(
y
,
p
)
]
{\displaystyle \displaystyle \Rightarrow {\frac {C}{p'}}=-{\frac {\partial }{\partial p}}[f(y,p)]}
(Eq. 5.7.24 )
Because the right hand side of (Eq. 5.7.24 ) is a function of two variables y and p, while the left hand side is a function of p' only, the equation (Eq. 5.7.24 ) could not hold if the constant
C
{\displaystyle C}
has a non-zero value. Thus, the only condition by which the equation (Eq. 5.7.24 ) will be satisfied is that
C
=
0
{\displaystyle C=0}
while
∂
∂
p
[
f
(
y
,
p
)
]
=
0
{\displaystyle {\frac {\partial }{\partial p}}[f(y,p)]=0}
,that is,
f
(
y
,
p
)
=
f
(
y
)
{\displaystyle f(y,p)=f(y)}
.
Substituting
C
=
0
{\displaystyle C=0}
into (Eq. 5.7.16 ) yields,
(
ϕ
p
x
−
ϕ
x
p
)
=
0
{\displaystyle \displaystyle (\phi _{px}-\phi _{xp})=0}
(Eq. 5.7.25 )
Thus we have
ϕ
p
x
=
ϕ
x
p
{\displaystyle \displaystyle \phi _{px}=\phi _{xp}}
(Eq. 5.7.26 )
We are now left with
(
ϕ
x
y
−
ϕ
y
x
)
=
0
{\displaystyle \displaystyle (\phi _{xy}-\phi _{yx})=0}
Thus
ϕ
x
y
=
ϕ
y
x
{\displaystyle \displaystyle \phi _{xy}=\phi _{yx}}
(Eq. 5.7.27 )
R*5.8. Working with the coefficients in 1st exactness condition Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
g
0
−
d
g
1
d
x
+
d
2
g
2
d
x
2
=
0
{\displaystyle \displaystyle g_{0}-{\frac {dg_{1}}{dx}}+{\frac {d^{2}g_{2}}{dx^{2}}}=0}
(Eq.(1) p.22-2 )
Using The Coefficients in the 1st exactness condition prove that (Eq.(1)p.22-3 ) can be written in the form
f
x
x
+
2
p
f
x
y
+
p
2
f
y
y
−
g
p
x
−
p
g
p
y
+
g
y
+
(
f
x
p
+
p
f
y
p
+
2
f
y
−
g
p
p
)
q
=
0
{\displaystyle \displaystyle f_{xx}+2pf_{xy}+p^{2}f_{yy}-g_{px}-pg_{py}+g_{y}+(f_{xp}+pf_{yp}+2f_{y}-g_{pp})q=0}
Nomenclature Edit
p
:=
y
(
x
)
′
{\displaystyle \displaystyle p:=y(x)'}
q
:=
y
(
x
)
″
{\displaystyle \displaystyle q:=y(x)''}
For an equation to be exact, they must be of the form
G
(
x
,
y
,
y
′
,
y
″
)
=
g
(
x
,
y
,
p
)
+
f
(
x
,
y
,
p
)
y
″
{\displaystyle \displaystyle G(x,y,y',y'')=g(x,y,p)+f(x,y,p)y''}
g
0
=
∂
G
∂
y
(
0
)
=
f
y
q
+
g
y
{\displaystyle \displaystyle g_{0}={\frac {\partial G}{\partial y^{(0)}}}=f_{y}q+g_{y}}
(Eq. 5.8.2 )
g
1
=
∂
G
∂
y
(
1
)
=
f
p
q
+
g
p
{\displaystyle \displaystyle g_{1}={\frac {\partial G}{\partial y^{(1)}}}=f_{p}q+g_{p}}
g
2
=
∂
G
∂
y
(
2
)
=
∂
G
∂
q
=
f
{\displaystyle \displaystyle g_{2}={\frac {\partial G}{\partial y^{(2)}}}={\frac {\partial G}{\partial q}}=f}
d
g
1
d
x
=
d
(
f
p
q
+
g
p
)
d
x
{\displaystyle \displaystyle {\frac {dg_{1}}{dx}}={\frac {d(f_{p}q+g_{p})}{dx}}}
using chain and product rule
d
g
1
d
x
=
f
x
p
q
+
f
y
p
p
q
+
f
p
p
q
2
+
g
p
x
+
g
p
y
p
+
g
p
p
q
{\displaystyle \displaystyle {\frac {dg_{1}}{dx}}=f_{xp}q+f_{yp}pq+f_{pp}q^{2}+g_{px}+g_{py}p+g_{pp}q}
(Eq. 5.8.3 )
d
g
2
d
x
=
d
(
f
(
x
,
y
,
p
)
)
d
x
{\displaystyle \displaystyle {\frac {dg_{2}}{dx}}={\frac {d(f(x,y,p))}{dx}}}
d
g
2
d
x
=
f
x
+
f
y
y
′
+
f
p
p
′
{\displaystyle \displaystyle {\frac {dg_{2}}{dx}}=f_{x}+f_{y}y'+f_{p}p'}
d
g
2
d
x
=
f
x
+
f
y
p
+
f
p
q
{\displaystyle \displaystyle {\frac {dg_{2}}{dx}}=f_{x}+f_{y}p+f_{p}q}
d
2
g
2
d
x
2
=
d
(
f
x
+
f
y
p
+
f
p
q
)
d
x
{\displaystyle \displaystyle {\frac {d^{2}g_{2}}{dx^{2}}}={\frac {d(f_{x}+f_{y}p+f_{p}q)}{dx}}}
d
2
g
2
d
x
2
=
f
x
x
+
f
x
y
p
+
f
y
x
q
+
f
y
x
p
+
f
y
y
p
2
+
f
y
p
q
p
+
f
x
p
q
+
f
p
y
p
q
+
f
p
p
q
2
{\displaystyle \displaystyle {\frac {d^{2}g_{2}}{dx^{2}}}=f_{xx}+f_{xy}p+f_{yx}q+f_{yx}p+f_{yy}p^{2}+f_{yp}qp+f_{xp}q+f_{py}pq+f_{pp}q^{2}}
(Eq. 5.8.4 )
plugging Eq(2),(3),&(40 into Eq(1)
f
y
q
+
g
y
−
[
f
x
p
q
+
f
y
p
p
q
+
f
p
p
q
2
+
g
p
x
+
g
p
y
p
+
g
p
p
q
]
+
f
x
x
+
f
x
y
p
+
f
y
x
q
+
f
y
x
p
+
f
y
y
p
2
+
f
y
p
q
p
+
f
x
p
q
+
f
p
y
p
q
+
f
p
p
q
2
=
0
{\displaystyle \displaystyle f_{y}q+g_{y}-[f_{xp}q+f_{yp}pq+f_{pp}q^{2}+g_{px}+g_{py}p+g_{pp}q]+f_{xx}+f_{xy}p+f_{yx}q+f_{yx}p+f_{yy}p^{2}+f_{yp}qp+f_{xp}q+f_{py}pq+f_{pp}q^{2}=0}
after cancellation of the opposite term
(
f
x
x
+
2
p
f
x
y
+
p
2
f
y
y
−
g
p
x
−
p
g
p
y
+
g
y
)
+
(
f
x
p
+
p
f
y
p
+
2
f
y
−
g
p
p
)
q
=
0
{\displaystyle \displaystyle (f_{xx}+2pf_{xy}+p^{2}f_{yy}-g_{px}-pg_{py}+g_{y})+(f_{xp}+pf_{yp}+2f_{y}-g_{pp})q=0}
Now, we can club the terms
(
f
x
x
+
2
p
f
x
y
+
p
2
f
y
y
−
g
p
x
−
p
g
p
y
+
g
y
)
1
=
g
¯
{\displaystyle \displaystyle (f_{xx}+2pf_{xy}+p^{2}f_{yy}-g_{px}-pg_{py}+g_{y})1={\bar {g}}}
and
(
f
x
p
+
p
f
y
p
+
2
f
y
−
g
p
p
)
q
=
f
¯
{\displaystyle \displaystyle (f_{xp}+pf_{yp}+2f_{y}-g_{pp})q={\bar {f}}}
Since 1 and q, i.e the second derivative of y, are in general non linear, for the equation to hold true, their coefficients must both be equal to zero.
Thus we say that
g
¯
=
(
f
x
x
+
2
p
f
x
y
+
p
2
f
y
y
−
g
p
x
−
p
g
p
y
+
g
y
)
1
=
0
{\displaystyle \displaystyle {\bar {g}}=(f_{xx}+2pf_{xy}+p^{2}f_{yy}-g_{px}-pg_{py}+g_{y})1=0}
and
f
¯
=
(
f
x
p
+
p
f
y
p
+
2
f
y
−
g
p
p
)
q
=
0
{\displaystyle \displaystyle {\bar {f}}=(f_{xp}+pf_{yp}+2f_{y}-g_{pp})q=0}
Which is the required proof.
R5.9: Use of MacLaurin Series Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
Use Taylor Series at x=0 (MacLaurin Series) to derive[13]
(
1
−
x
)
−
a
=
1
+
a
x
+
a
(
a
+
1
)
x
2
2
!
+
a
(
a
+
1
)
(
a
+
2
)
x
3
3
!
+
.
.
.
.
.
.
.
=
F
(
a
,
b
;
b
;
x
)
{\displaystyle \displaystyle (1-x)^{-a}=1+ax+a(a+1){\frac {x^{2}}{2!}}+a(a+1)(a+2){\frac {x^{3}}{3!}}+.......=F(a,b;b;x)}
1
x
a
r
c
t
a
n
(
1
+
x
)
=
1
−
x
2
3
+
x
4
5
−
x
6
7
+
.
.
.
.
=
F
(
1
2
,
1
,
3
2
,
−
x
2
)
{\displaystyle \displaystyle {\frac {1}{x}}arctan(1+x)=1-{\frac {x^{2}}{3}}+{\frac {x^{4}}{5}}-{\frac {x^{6}}{7}}+....=F\left({\frac {1}{2}},1,{\frac {3}{2}},-x^{2}\right)}
The Taylor's series [14] expansion of a function f(x) about a real or complex number c is given by the formula
f
(
c
)
+
f
′
(
c
)
1
!
(
x
−
c
)
+
f
″
(
c
)
2
!
(
x
−
c
)
2
+
f
(
3
)
(
c
)
3
!
(
x
−
c
)
3
+
⋯
.
{\displaystyle f(c)+{\frac {f'(c)}{1!}}(x-c)+{\frac {f''(c)}{2!}}(x-c)^{2}+{\frac {f^{(3)}(c)}{3!}}(x-c)^{3}+\cdots .}
(Eq. 5.9.1 )
When the neighborhood for the expansion is zero, i.e c = 0, the resulting series is called the Maclaurin Series.
We have the function
f
(
x
)
=
1
(
1
−
x
)
a
{\displaystyle \displaystyle f(x)={\frac {1}{(1-x)^{a}}}}
Rewriting the Maclaurin series expansion,
1
(
1
−
x
)
a
=
f
(
0
)
+
f
′
(
0
)
1
!
(
x
−
0
)
+
f
″
(
0
)
2
!
(
x
−
0
)
2
+
f
(
3
)
(
0
)
3
!
(
x
−
0
)
3
+
⋯
.
{\displaystyle \displaystyle {\frac {1}{(1-x)^{a}}}=f(0)+{\frac {f'(0)}{1!}}(x-0)+{\frac {f''(0)}{2!}}(x-0)^{2}+{\frac {f^{(3)}(0)}{3!}}(x-0)^{3}+\cdots .}
(Eq. 5.9.2 )
Substituting the values from the tables in (Eq. 5.9.2 ) we get
1
(
1
−
x
)
a
=
1
+
a
1
!
(
x
)
+
a
(
a
+
1
)
2
!
(
x
)
2
+
a
(
a
+
1
)
(
a
+
2
)
3
!
(
x
)
3
+
⋯
.
{\displaystyle \displaystyle {\frac {1}{(1-x)^{a}}}=1+{\frac {a}{1!}}(x)+{\frac {a(a+1)}{2!}}(x)^{2}+{\frac {a(a+1)(a+2)}{3!}}(x)^{3}+\cdots .}
(Eq. 5.9.3 )
1
(
1
−
x
)
a
=
∑
n
=
0
∞
(
a
)
k
x
k
n
!
{\displaystyle \displaystyle {\frac {1}{(1-x)^{a}}}=\sum _{n=0}^{\infty }(a)_{k}{\frac {x^{k}}{n!}}}
(Eq. 5.9.4 )
Where[15]
(
a
)
0
:=
1
{\displaystyle (a)_{0}:=1}
(
a
)
k
:=
a
(
a
+
1
)
(
a
+
2
)
⋯
(
a
+
k
−
1
)
{\displaystyle (a)_{k}:=a(a+1)(a+2)\cdots (a+k-1)}
We can represent
∑
n
=
0
∞
(
a
)
k
(
b
)
k
(
c
)
k
x
k
n
!
=
F
(
a
,
b
;
c
;
x
)
{\displaystyle \sum _{n=0}^{\infty }{\frac {(a)_{k}\,(b)_{k}}{(c)_{k}}}\,{\frac {x^{k}}{n!}}=F(a,b;c;x)}
(Eq. 5.9.4 ) can be written as
∑
n
=
0
∞
(
a
)
k
(
b
)
k
(
b
)
k
x
k
n
!
=
F
(
a
,
b
;
b
;
x
)
{\displaystyle \sum _{n=0}^{\infty }{\frac {(a)_{k}(b)_{k}}{(b)_{k}}}{\frac {x^{k}}{n!}}=F(a,b;b;x)}
, hence proved.
We have the function
1
x
a
r
c
t
a
n
(
1
+
x
)
{\displaystyle \displaystyle {\frac {1}{x}}arctan(1+x)}
We will use a slightly different approach here when compared to part a of the solution. We will expand
a
r
c
t
a
n
(
1
+
x
)
{\displaystyle \displaystyle arctan(1+x)}
and multiply the resulting expanded function with
1
x
{\displaystyle \displaystyle {\frac {1}{x}}}
Rewriting the Maclaurin series expansion,
a
r
c
t
a
n
(
1
+
x
)
=
f
(
0
)
+
f
′
(
0
)
1
!
(
x
−
0
)
+
f
″
(
0
)
2
!
(
x
−
0
)
2
+
f
(
3
)
(
0
)
3
!
(
x
−
0
)
3
+
⋯
.
{\displaystyle \displaystyle arctan(1+x)=f(0)+{\frac {f'(0)}{1!}}(x-0)+{\frac {f''(0)}{2!}}(x-0)^{2}+{\frac {f^{(3)}(0)}{3!}}(x-0)^{3}+\cdots .}
(Eq. 5.9.5 )
Substituting the values from the tables in (Eq. 5.9.5 ) we get
a
r
c
t
a
n
(
1
+
x
)
=
π
4
+
1
2
∗
1
!
(
x
)
−
1
2
∗
2
!
(
x
)
2
+
1
2
∗
3
!
(
x
)
3
+
⋯
.
{\displaystyle \displaystyle arctan(1+x)={\frac {\pi }{4}}+{\frac {1}{2*1!}}(x)-{\frac {1}{2*2!}}(x)^{2}+{\frac {1}{2*3!}}(x)^{3}+\cdots .}
(Eq. 5.9.6 )
Multiplying (Eq. 5.9.6 ) with
1
x
{\displaystyle \displaystyle {\frac {1}{x}}}
1
x
a
r
c
t
a
n
(
1
+
x
)
=
π
4
x
+
1
2
∗
1
!
−
1
2
∗
2
!
(
x
)
+
1
2
∗
3
!
(
x
)
2
+
⋯
.
{\displaystyle \displaystyle {\frac {1}{x}}arctan(1+x)={\frac {\pi }{4x}}+{\frac {1}{2*1!}}-{\frac {1}{2*2!}}(x)+{\frac {1}{2*3!}}(x)^{2}+\cdots .}
This expression does not match the expression that we have been asked to prove. This, we believe is because there has been a misprint and the expression to be found out must be
1
x
a
r
c
t
a
n
(
x
)
{\displaystyle {\frac {1}{x}}arctan(x)}
Expanding
a
r
c
t
a
n
(
x
)
{\displaystyle arctan(x)}
using Maclaurin's series
Rewriting the Maclaurin series expansion,
a
r
c
t
a
n
(
x
)
=
f
(
0
)
+
f
′
(
0
)
1
!
(
x
−
0
)
+
f
″
(
0
)
2
!
(
x
−
0
)
2
+
f
(
3
)
(
0
)
3
!
(
x
−
0
)
3
+
⋯
.
{\displaystyle \displaystyle arctan(x)=f(0)+{\frac {f'(0)}{1!}}(x-0)+{\frac {f''(0)}{2!}}(x-0)^{2}+{\frac {f^{(3)}(0)}{3!}}(x-0)^{3}+\cdots .}
(Eq. 5.9.7 )
Substituting the values from the tables in (Eq. 5.9.7 ) we get
a
r
c
t
a
n
(
x
)
=
0
+
1
1
!
(
x
)
+
0
2
!
(
x
)
2
+
−
2
3
!
(
x
)
3
+
0
4
!
(
x
)
4
+
24
5
!
(
x
)
5
⋯
.
{\displaystyle \displaystyle arctan(x)=0+{\frac {1}{1!}}(x)+{\frac {0}{2!}}(x)^{2}+{\frac {-2}{3!}}(x)^{3}+{\frac {0}{4!}}(x)^{4}+{\frac {24}{5!}}(x)^{5}\cdots .}
(Eq. 5.9.8 )
Multiplying (Eq. 5.9.8 ) with
1
x
{\displaystyle \displaystyle {\frac {1}{x}}}
a
r
c
t
a
n
(
x
)
=
0
+
1
1
!
+
−
2
3
!
(
x
)
2
+
24
5
!
(
x
)
4
⋯
.
{\displaystyle \displaystyle arctan(x)=0+{\frac {1}{1!}}+{\frac {-2}{3!}}(x)^{2}+{\frac {24}{5!}}(x)^{4}\cdots .}
a
r
c
t
a
n
(
x
)
=
1
−
(
x
)
2
3
+
(
x
)
4
5
+
(
x
)
6
7
⋯
.
{\displaystyle \displaystyle arctan(x)=1-{\frac {(x)^{2}}{3}}+{\frac {(x)^{4}}{5}}+{\frac {(x)^{6}}{7}}\cdots .}
(Eq. 5.9.9 )
Which is the expression in the RHS.
R5.10 Gauss Hypergeometric Series [16] Edit
On our honor, we did this problem on our own, without looking at the solutions in previous semesters or other online solutions.
1. Use MATLAB to plot
F
(
5
,
−
10
;
1
;
x
)
{\displaystyle F(5,-10;1;x)}
near x=0 to show the local maximum (or maxima) in this region.
2. Show that
F
(
5
,
−
10
;
1
;
x
)
=
(
1
−
x
)
6
(
1001
x
4
−
1144
x
3
+
396
x
2
−
44
x
+
1
)
{\displaystyle F(5,-10;1;x)=(1-x)^{6}(1001x^{4}-1144x^{3}+396x^{2}-44x+1)}
((1) pg. 64-9b )
The MATLAB code, shown below, will plot the hypergeometric function
F
(
5
,
−
10
;
1
;
x
)
{\displaystyle F(5,-10;1;x)}
over the interval:
0
≤
x
≤
0.8
{\displaystyle 0\leq x\leq 0.8}
.
x = [ 0 : 0.01 : 0.8 ] ' ;
plot ( x , hypergeom ([ 5 , - 10 ], 1 , x ))
The plot of the hypergeometric function near x=0 reveals a local maximum of 0.1481 at x = 0.23.
The hypergeometric function
F
(
5
,
−
10
;
1
;
x
)
{\displaystyle F(5,-10;1;x)}
can be expressed as
∑
k
=
0
∞
(
a
)
k
(
b
)
k
(
c
)
k
x
k
k
!
{\displaystyle \sum _{k=0}^{\infty }{\frac {(a)_{k}(b)_{k}}{(c)_{k}}}{\frac {x^{k}}{k!}}}
using the Pochhammer Symbol
where
(
a
)
k
(
b
)
k
(
c
)
k
=
a
(
a
+
1
)
(
a
+
2
)
.
.
.
(
a
+
k
−
1
)
b
(
b
+
1
)
(
b
+
2
)
.
.
.
(
b
+
k
−
1
)
c
(
c
+
1
)
(