An th-rank
tensor in
-dimensional space is a mathematical
object that has
indices and
components and obeys certain transformation rules. Each index
of a tensor ranges over the number of dimensions of space.
However, the dimension of the space is largely irrelevant in most tensor equations
(with the notable exception of the contracted Kronecker
delta). Tensors are generalizations of scalars (that have no indices), vectors
(that have exactly one index), and matrices (that
have exactly two indices) to an arbitrary number of indices.
Tensors provide a natural and concise mathematical framework for formulating and solving problems in areas of physics such as elasticity, fluid mechanics, and general relativity.
The notation for a tensor is similar to that of a matrix (i.e., ), except that a tensor
,
,
, etc., may have an arbitrary number of indices.
In addition, a tensor with rank
may be of mixed type
, consisting of
so-called "contravariant" (upper) indices
and
"covariant" (lower) indices.
Note that the positions of the slots in which contravariant and covariant indices
are placed are significant so, for example,
is distinct from
.
While the distinction between covariant and contravariant indices must be made for general tensors, the two are equivalent for tensors in three-dimensional Euclidean space, and such tensors are known as Cartesian tensors.
Objects that transform like zeroth-rank tensors are called scalars, those that transform like first-rank tensors are called vectors,
and those that transform like second-rank tensors are called matrices.
In tensor notation, a vector
would be written
,
where
, ...,
, and matrix is a tensor of type
, which would be written
in tensor notation.
Tensors may be operated on by other tensors (such as metric tensors, the permutation tensor, or the
Kronecker delta) or by tensor operators (such
as the covariant derivative). The manipulation
of tensor indices to produce identities or to simplify expressions is known as index gymnastics, which includes index
lowering and index raising as special cases.
These can be achieved through multiplication by a so-called metric
tensor ,
,
, etc., e.g.,
(1)
| |||
(2)
|
(Arfken 1985, p. 159).
Tensor notation can provide a very concise way of writing vector and more general identities. For example, in tensor notation, the dot product is simply written
(3)
|
where repeated indices are summed over (Einstein summation). Similarly, the cross product can be concisely written as
(4)
|
where is the permutation
tensor.
Contravariant second-rank tensors are objects which transform as
(5)
|
Covariant second-rank tensors are objects which transform as
(6)
|
Mixed second-rank tensors are objects which transform as
(7)
|
If two tensors
and
have the same rank and the same covariant
and contravariant indices, then they can
be added in the obvious way,
(8)
| |||
(9)
| |||
(10)
|
The generalization of the dot product applied to tensors is called tensor contraction, and consists of setting two unlike indices equal to each other and then summing using the Einstein summation convention. Various types of derivatives can be taken of tensors, the most common being the comma derivative and covariant derivative.
If the components of any tensor of any tensor rank vanish in one particular coordinate system, they vanish in all coordinate systems. A transformation of the variables of a tensor changes the tensor into another whose components are linear homogeneous functions of the components of the original tensor.
A tensor space of type can be described as a vector
space tensor product between
copies of vector fields and
copies of the dual vector fields, i.e.,
one-forms. For example,
(11)
|
is the vector bundle of -tensors on a manifold
, where
is the tangent bundle of
and
is its dual. Tensors of type
form a vector space. This
description generalized to any tensor type, and an invertible
linear map
induces a map
,
where
is the dual
vector space and
the Jacobian, defined by
(12)
|
where is the pullback
map of a form is defined using the transpose of the Jacobian.
This definition can be extended similarly to other tensor
products of
and
. When there is a change of coordinates,
then tensors transform similarly, with
the Jacobian of the linear transformation.