Search results
Results from the WOW.Com Content Network
As the name implies, the divergence is a (local) measure of the degree to which vectors in the field diverge. The divergence of a tensor field of non-zero order k is written as =, a contraction of a tensor field of order k − 1. Specifically, the divergence of a vector is a scalar.
In vector calculus, divergence is a vector operator that operates on a vector field, producing a scalar field giving the quantity of the vector field's source at each point. More technically, the divergence represents the volume density of the outward flux of a vector field from an infinitesimal volume around a given point.
The gradient of F is then normal to the hypersurface. Similarly, an affine algebraic hypersurface may be defined by an equation F(x 1, ..., x n) = 0, where F is a polynomial. The gradient of F is zero at a singular point of the hypersurface (this is the definition of a singular point). At a non-singular point, it is a nonzero normal vector.
Del is a very convenient mathematical notation for those three operations (gradient, divergence, and curl) that makes many equations easier to write and remember. The del symbol (or nabla) can be formally defined as a vector operator whose components are the corresponding partial derivative operators.
Vector calculus or vector analysis is a branch of mathematics concerned with the differentiation and integration of vector fields, primarily in three-dimensional Euclidean space, . [1] The term vector calculus is sometimes used as a synonym for the broader subject of multivariable calculus, which spans vector calculus as well as partial differentiation and multiple integration.
More precisely, the divergence theorem states that the surface integral of a vector field over a closed surface, which is called the "flux" through the surface, is equal to the volume integral of the divergence over the region enclosed by the surface. Intuitively, it states that "the sum of all sources of the field in a region (with sinks ...
The divergence of a tensor field () is defined using the recursive relation = ; = () where c is an arbitrary constant vector and v is a vector field. If T {\displaystyle {\boldsymbol {T}}} is a tensor field of order n > 1 then the divergence of the field is a tensor of order n − 1.
The gradient theorem states that if the vector field F is the gradient of some scalar-valued function (i.e., if F is conservative), then F is a path-independent vector field (i.e., the integral of F over some piecewise-differentiable curve is dependent only on end points). This theorem has a powerful converse: