Here we extend the concept of vector to that of the vector field. A familiar example of a vector field is wind velocity: It has direction and magnitude, which makes it a vector. But it also depends on position (and ultimately on time). Wind velocity is a function of (x,y,z) at any given time, equivalently we can say that wind velocity is a time-dependent field: .
Let be a scalar function. Assume that the partial derivatives of the function are continuous in some region of space. If the point has coordinates () with respect to the basis (), the gradient of is defined as
In index notation,
The gradient is obviously a vector and has a direction. We can think of the gradient at a point being the vector perpendicular to the level contour at that point.
It is often useful to think of the symbol as an operator of the form
If we form a scalar product of a vector field with the operator, we get a scalar quantity called the
divergence of the vector field. Thus,
In index notation,
If , then is called a divergence-free field.
The physical significance of the divergence of a vector field is the rate at which some density exits a given region of space. In the absence of the creation or destruction of matter, the density within a region of space can change only by having it flow into or out of the region.
One version of the fundamental theorem of one-dimensional calculus is
This is a theorem about a function, , its first derivative, and a line segment. Two notations used to denote this line segment are [a,b] and the inequality, a<x<b. In the field of topology, denotes boundary. If we let the symbol denote the infinite number of points in the line segment [a,b], then the symbol denotes the two endpoints (at x = a and x = b ) of the line segment . These endpoints form the boundary of .
The gradient theorem is a direct generalization of the fundamental theorem of calculus:
The subscript, informs this is an integral over the over a one-dimensinal curve (or 'path') line integral from point to point . The function, is any scalar field that is differentiable. The expression informs us that can be a member of an n-dimensional space. (In other words the theorem is easily generalized to more than three dimensions.) A consequence of this theorem is that for any "closed curve" The figure shows the closed curve A, as well as the "open curve", B. Two endpoints form the "boundary" of curve B.
The integral subscript, informs us that this theorem is valid only in a three-dimensional vector space. The integral is over a two-dimensional surface,Σ ,with , where is normal to the surface. The integral over the surface, Σ, is nonzero only if its boundary, ∂Σ, exists. Surfaces with such boundaries are called open surfaces, and the boundary, ∂Σ, is a curve in 3-space that goes along the "edge" of the surface. This curve is integrated in the direction of positive orientation, meaning that and the surface normal follow follow the right-hand rule.
Footnote: According to Wikipedia[2], this form of the theorem was first discovered by Lord Kelvin, who communicated it to George Stokes in a letter dated July 2, 1850. Stokes set the theorem as a question on the 1854 Smith's Prize exam, which led to the result bearing his name.
The integral subscript, , informs us that this theorem is valid in an (arbitrary) n-dimensional vector space. The n-dimensional volume is Ω, and ∂Ω is its boundary. If n =3 dimensions, ∂Ω is a surface. Since this surface encloses a volume, it has no boundary of its own, and is therefore called a closed surface. The figure shows six surfaces. The three on the left have no boundary and are therefore closed; the ones to the right have a boundary (shown in red) and are therefore open. Note that the closed surfaces to the left are themselves boundaries volumes which are defined as what is "inside" the surface.
Footnote: In index notation, the gradient theorem can be written as