Complex Analysis/Partial derivative

Introduction

edit

This page about Complex Analysis/Partial derivative can be displayed as Wiki2Reveal slides. Single sections are regarded as slides and modifications on the slides will immediately affect the content of the slides. The following aspects of Complex Analysis/Partial derivative are considered in detail:

  • (1) Partial derivative in real numbers
  • (2) Cauchy-Riemann equations

Objective

edit

This learning resource about Partial Derivatives in Wikiversity has the objective to link Calculus and Complex Analysis.

Topic - Calculus and Complex Analysis

edit

In mathematics, a partial derivative of a function of several variables is its derivative with respect to one of those variables, with the others held constant (as opposed to the total derivative, in which all variables are allowed to vary). Partial derivatives are used in vector calculus, differential geometry and in Cauchy-Riemann equations.

Denotation 1

edit

The partial derivative of a function   with respect to the variable   is variously denoted by

 ,  ,  ,  ,  ,  , or  .

It can be thought of as the rate of change of the function in the  -direction.

Denotation 2

edit

Sometimes, for  , the partial derivative of   with respect to   is denoted as   Since a partial derivative generally has the same arguments as the original function, its functional dependence is sometimes explicitly signified by the notation, such as in:

 

Symbol

edit

The symbol used to denote partial derivatives is . One of the first known uses of this symbol in mathematics is by Marquis de Condorcet from 1770,[1] who used it for partial differences. The modern partial derivative notation was created by Adrien-Marie Legendre (1786), although he later abandoned it; Carl Gustav Jacob Jacobi reintroduced the symbol in 1841.[2]

Definition

edit

Like ordinary derivatives, the partial derivative is defined as a limit. Let   be an open subset of   and   a function. The partial derivative of   at the point   with respect to the  -th variable   is defined as

 

Partial Derivation - Unit Vector

edit

Where   is the unit vector of  -th variable  . Even if all partial derivatives   exist at a given point  , the function need not be continuous there.

Total Derivative

edit

However, if all partial derivatives exist in a neighborhood of   and are continuous there, then   is totally differentiable in that neighborhood and the total derivative is continuous. In this case, it is said that   is a   function. This can be used to generalize for vector valued functions, {{nowrap|  by carefully using a componentwise argument.

Partial Derivative of higher Order

edit

The partial derivative   can be seen as another function defined on   and can again be partially differentiated. If the direction of derivative is not repeated, it is called a mixed partial derivative. If all mixed second order partial derivatives are continuous at a point (or on a set),   is termed a   function at that point (or on that set); in this case, the partial derivatives can be exchanged by Clairaut's theorem:

 

Notation - Order of Derivative

edit
see ∂ symbol

For the following examples, let   be a function in    , and  .

First-order partial derivatives

edit

 

Second-order partial derivatives

edit

 

Second-order mixed derivatives:

 

Higher-order partial and mixed derivatives

edit

Higher-order partial and mixed derivatives are denoted.

 

Derivative - Variables

edit

When dealing with functions of multiple variables it is necessary to specify explicitly for which variables the partial derivation are performed and how often are being held constant to. In fields such as statistical mechanics, the partial derivative of   with respect to  , holding   and   constant, is often expressed as

 

Conventionally, for clarity and simplicity of notation, the partial derivative function and the value of the function at a specific point are conflated by including the function arguments when the partial derivative symbol (Leibniz notation) is used. Thus, an expression like

 

is used for the function, while

 

might be used for the value of the function at the point  . However, this convention breaks down when we want to evaluate the partial derivative at a point like  . In such a case, evaluation of the function must be expressed in an unwieldy manner as

 

or

 

in order to use the Leibniz notation. Thus, in these cases, it may be preferable to use the Euler differential operator notation with   as the partial derivative symbol with respect to the  -th variable. For instance, one would write   for the example described above, while the expression   represents the partial derivative function with respect to the first variable.[3]

For higher order partial derivatives, the partial derivative (function) of   with respect to the  -th variable is denoted  . That is,  , so that the variables are listed in the order in which the derivatives are taken, and thus, in reverse order of how the composition of operators is usually notated. Of course, Clairaut's theorem implies that   as long as comparatively mild regularity conditions on   are satisfied.

Gradient

edit

An important example of a function of several variables is the case of a scalar-valued function   on a domain in Euclidean space   (e.g., on   or  ). In this case   has a partial derivative   with respect to each variable  . At the point  , these partial derivatives define the vector

 

This vector is called the gradient of   at  . If   is differentiable at every point in some domain, then the gradient is a vector-valued function   which takes the point   to the vector  . Consequently, the gradient produces a vector field.

A common abuse of notation is to define the del operator ( ) as follows in three-dimensional Euclidean space   with unit vectors  :

 

Or, more generally, for  -dimensional Euclidean space   with coordinates   and unit vectors  :

 

Example

edit

Suppose that   is a function of more than one variable. For instance,

 

 
A graph of  . For the partial derivative at (1, 1) that leaves   constant, the corresponding tangent line is parallel to the  -plane.
A slice of the graph above showing the function in the  -plane at  . The two axes are shown here with different scales. The slope of the tangent line is 3.

The graph of this function defines a surface in Euclidean space. To every point on this surface, there are an infinite number of tangent lines. Partial differentiation is the act of choosing one of these lines and finding its slope. Usually, the lines of most interest are those that are parallel to the  -plane, and those that are parallel to the  -plane (which result from holding either   or   constant, respectively).

To find the slope of the line tangent to the function at   and parallel to the  -plane, we treat   as a constant. The graph and this plane are shown on the right. Below, we see how the function looks on the plane  . By finding the derivative of the equation while assuming that   is a constant, we find that the slope of   at the point   is:

 

So at  , by substitution, the slope is  . Therefore,

 

at the point  . That is, the partial derivative of   with respect to   at   is  , as shown in the graph.

The function   can be reinterpreted as a family of functions of one variable indexed by the other variables:

 

In other words, every value of   defines a function, denoted  , which is a function of one variable  .[4] That is,

 

In this section the subscript notation   denotes a function contingent on a fixed value of  , and not a partial derivative.

Once a value of   is chosen, say  , then   determines a function   which traces a curve   on the  -plane:

 

In this expression,   is a constant, not a variable, so   is a function of only one real variable, that being  . Consequently, the definition of the derivative for a function of one variable applies:

 

The above procedure can be performed for any choice of  . Assembling the derivatives together into a function gives a function which describes the variation of   in the   direction:

 

This is the partial derivative of   with respect to  . Here   is a rounded d called the partial derivative symbol; to distinguish it from the letter 'd', ' ' is sometimes pronounced "partial".

Higher order partial derivatives

edit

Second and higher order partial derivatives are defined analogously to the higher order derivatives of univariate functions. For multivariate function   it has to be specified for which second partial derivative with respect to   is simply the partial derivative of the partial derivative (both with respect to  ):[5]:316–318

 

The cross partial derivative with respect to   and   is obtained by taking the partial derivative of   with respect to  , and then taking the partial derivative of the result with respect to  , to obtain

 

Schwarz's theorem states that if the second derivatives are continuous, the expression for the cross partial derivative is unaffected by which variable the partial derivative is taken with respect to first and which is taken second. That is,

 

or equivalently  

Own and cross partial derivatives appear in the Hessian matrix which is used in the second order conditions in optimization problems. The higher order partial derivatives can be obtained by successive differentiation

Antiderivative analogue

edit

There is a concept for partial derivatives that is analogous to antiderivatives for regular derivatives. Given a partial derivative, it allows for the partial recovery of the original function.

Consider the example of

 

The so-called partial integral can be taken with respect to   (treating   as constant, in a similar manner to partial differentiation):

 

Here, the constant of integration is no longer a constant, but instead a function of all the variables of the original function except  . The reason for this is that all the other variables are treated as constant when taking the partial derivative, so any function which does not involve   will disappear when taking the partial derivative, and we have to account for this when we take the antiderivative. The most general way to represent this is to have the constant represent an unknown function of all the other variables.

Thus the set of functions  , where   is any one-argument function, represents the entire set of functions in variables   that could have produced the  -partial derivative  .

If all the partial derivatives of a function are known (for example, with the gradient), then the antiderivatives can be matched via the above process to reconstruct the original function up to a constant. Unlike in the single-variable case, however, not every set of functions can be the set of all (first) partial derivatives of a single function. In other words, not every vector field is conservative.

Applications

edit

Geometry

edit
 
The volume of a cone depends on height and radius

The volume   of a cone depends on the cone's height   and its radius   according to the formula

 

The partial derivative of   with respect to   is

 

which represents the rate with which a cone's volume changes if its radius is varied and its height is kept constant. The partial derivative with respect to   equals  , which represents the rate with which the volume changes if its height is varied and its radius is kept constant.

By contrast, the total derivative of   with respect to   and   are respectively

 

The difference between the total and partial derivative is the elimination of indirect dependencies between variables in partial derivatives.

If (for some arbitrary reason) the cone's proportions have to stay the same, and the height and radius are in a fixed ratio  ,

 

This gives the total derivative with respect to  ,

 

which simplifies to

 

Similarly, the total derivative with respect to   is

 

The total derivative with respect to both   and   of the volume intended as scalar function of these two variables is given by the gradient vector

 

Optimization

edit

Partial derivatives appear in any calculus-based optimization problem with more than one choice variable. For example, in economics a firm may wish to maximize profit   with respect to the choice of the quantities   and   of two different types of output. The first order conditions for this optimization are  . Since both partial derivatives   and   will generally themselves be functions of both arguments   and  , these two first order conditions form a system of two equations in two unknowns.

Thermodynamics, quantum mechanics and mathematical physics

edit

Partial derivatives appear in thermodynamic equations like Gibbs-Duhem equation, in quantum mechanics as in Schrödinger wave equation, as well as in other equations from mathematical physics. The variables being held constant in partial derivatives here can be ratios of simple variables like mole fractions   in the following example involving the Gibbs energies in a ternary mixture system:

 

Mole Fraction

edit

Express mole fractions of a component as functions of other components' mole fraction and binary mole ratios:

 


Differential Quotients

edit

Differential quotients can be formed at constant ratios like those above:

 

Multicomponent Systems

edit

Ratios   of mole fractions can be written for ternary and multicomponent systems:

 

Solving Partial Differential Equations

edit

The ratios above can be used for solving partial differential equations like:

 

This equality can be rearranged to have differential quotient of mole fractions on one side.

Image resizing

edit

Partial derivatives are key to target-aware image resizing algorithms. Widely known as seam carving, these algorithms require each pixel in an image to be assigned a numerical 'energy' to describe their dissimilarity against orthogonal adjacent pixels. The algorithm then progressively removes rows or columns with the lowest energy. The formula established to determine a pixel's energy (magnitude of gradient at a pixel) depends heavily on the constructs of partial derivatives.

Economics

edit

Partial derivatives play a prominent role in economics, in which most functions describing economic behaviour posit that the behaviour depends on more than one variable. For example, a societal consumption function may describe the amount spent on consumer goods as depending on both income and wealth; the marginal propensity to consume is then the partial derivative of the consumption function with respect to income.

See also

edit

References

edit
  1. Cajori, Florian (1952), A History of Mathematical Notations, vol. 2 (3 ed.), 596
  2. Miller, Jeff (n.d.). "Earliest Uses of Symbols of Calculus". In O'Connor, John J.; Robertson, Edmund F. (eds.). MacTutor History of Mathematics archive. University of St Andrews. Retrieved 2023-06-15.
  3. Spivak, M. (1965). Calculus on Manifolds. New York: W. A. Benjamin. pp. 44. ISBN 9780805390216. https://archive.org/details/SpivakM.CalculusOnManifoldsPerseus2006Reprint. 
  4. This can also be expressed as the adjointness between the product space and function space constructions.
  5. Chiang, Alpha C. (1984). Fundamental Methods of Mathematical Economics (3rd ed.). McGraw-Hill. 


Seiteninformation

edit

Diese Lernresource können Sie als Wiki2Reveal-Foliensatz darstellen.

Wiki2Reveal

edit

Dieser Wiki2Reveal Foliensatz wurde für den Lerneinheit Complex Analysis' erstellt der Link für die Wiki2Reveal-Folien wurde mit dem Wiki2Reveal-Linkgenerator erstellt.

Page Information

edit

Wikipedia2Wikiversity

edit

This page was based on the following wikipedia-source page: