It is common practice to use matrices to represent transformations of a vector into another vector. Here, we discuss another quantity, known as a tensor, that achieves the same purpose. We generally denote tensors by uppercase boldfaced symbols, such as , and symbolize the transformation of a vector
by
to a vector
as
(1)
The advantages of using tensors are that they are often far more compact than matrices, they are easier to differentiate, and their components transform transparently under changes of bases. The latter feature is very useful when interpreting results in rigid-body kinematics. Consequently, we choose to employ a tensor notation throughout this site for many of the developments we present. This page is intended as both a brief introduction to and review of tensors and their properties. The material provided in the forthcoming sections is standard background for courses in continuum mechanics, and, as such, the primary sources of this information are works by Casey [1, 2], Chadwick [3], and Gurtin [4].
Contents
Preliminaries
To lay the foundation for our upcoming exposition on tensors, we begin with a brief discussion of basis vectors and define two symbols that prove useful in subsequent sections.
Basis vectors
Euclidean three-space is denoted by . For this space, we define a fixed, right-handed orthonormal basis
. By orthonormal, we mean that for any set of vectors
, the dot products
when
and
when
. Unless indicated otherwise, lowercase italic Latin indices such as
,
, and
range from 1 to 3. If
is right-handed, then the following scalar triple product is positive:
(2)
We also make use of another right-handed orthonormal basis, , that is not necessarily fixed.
The Kronecker delta
We use copious amounts of dot products, so it is convenient to define the Kronecker delta :
(3)
Clearly,
(4)
The alternating symbol
We also occasionally make use of the alternating (or Levi-Civita) symbol , which is defined such that
(5)
In words, if
is an even permutation of 1, 2, and 3;
if
is an odd permutation of 1, 2, and 3; and
if either
,
, or
. We also note that
(6)
which is simple to verify by using the definition of the scalar triple product.
The tensor product of two vectors
The tensor (or cross-bun) product of any two vectors and
in
is defined by
(7)
where is any vector in
. That is,
projects
onto
and multiplies the resulting scalar by
. Put another way,
transforms
into a vector that is parallel to
. A related tensor product is defined as follows:
(8)
In either case, performs a linear transformation of
that it acts on. The tensor product has the following useful properties:
(9)
where and
are any two scalars. To prove these identities, one merely shows that the left- and right-hand sides provide the same transformation of any vector
in
.
Second-order tensors
A second-order tensor is a linear transformation of
into itself. That is, for any two vectors
and
and any two scalars
and
,
(10)
where and
are both vectors in
. The tensor
is a simple example of a second-order tensor. It is standard to define the following composition rules for second-order tensors:
(11)
where and
are any second-order tensors. To check if two second-order tensors
and
are identical, it suffices to show that
for any
. We also define the identity tensor
and the zero tensor
:
(12)
Representations
It is convenient at this stage to establish the following representation for any second-order tensor :
(13)
where
(14)
are the components of relative to the basis
. The order of the indices
and
is important. Initially, it is convenient to interpret a tensor using the representation
(15)
for which
(16)
In this light, transforms
into
. Hence, if we know what
does to three orthonormal vectors, then we can write its representation immediately. To arrive at representation (13), we examine the action of a second-order tensor
on any vector
:
(17)
where we used the definition of the tensor product of two vectors in the next-to-last step. Thus, we infer that has the representation given by (13). We can use this representation to establish expressions for the transformation induced by
. To proceed, define
, in which case, from (17),
(18)
The components of
are then given by
(19)
When expressed in matrix notation, (19) has a familiar form:
(20)
Note that (20) implies that the identity tensor has the representation .
The product of two second-order tensors
We now turn to the product of two second-order tensors and
. The product
is defined here to be a second-order tensor
. First, let
(21)
We then solve the equations
(22)
for the nine components of , where
is any vector. Using the arbitrariness of
, we conclude that the components of the three tensors, which are all expressed in the same basis, are related by
(23)
This result is identical to that for matrix multiplication. Indeed, if we define three matrices whose components are ,
, and
, then we find the representation
(24)
It is straightforward to establish a similar representation for the product . Finally, consider the product of two second-order tensors
and
:
(25)
This result is the simplest way to remember how to multiply two second-order tensors.
Symmetric and skew-symmetric tensors
The transpose of a second-order tensor
is defined such that
(26)
for any two vectors and
. If we consider the second-order tensor
, then we can use definition (26) to show that
(27)
Given any two second-order tensors and
, it can be shown that the transpose
. If
, then
is said to be symmetric. On the other hand,
is skew-symmetric if
. Using the representation (13) for
and the identity (27)1, we find that the tensor components
when
is symmetric and
when
is skew-symmetric. These results imply that
has six independent components when it is symmetric but only three independent components when skew-symmetric. Lastly, it is always possible to decompose any second-order tensor
into the sum of a symmetric tensor and a skew-symmetric tensor:
(28)
Invariants
There are three scalar quantities associated with a second-order tensor that are independent of the right-handed orthonormal basis used for . Because these quantities are independent of the basis, they are known as the (principal) invariants of a second-order tensor. Given a second-order tensor
, the invariants
,
, and
of
are defined as
(29)
where ,
, and
are any three vectors. The first invariant
is known as the trace of a tensor
, and the third invariant
is known as the determinant of
:
(30)
Suppose we represent in terms of a right-handed orthonormal basis
in
, such as in representation (13). If we take
,
, and
, then
(31)
Consequently, the trace of is given by
(32)
A similar result holds for the trace of a matrix. We also note the related result . In addition, we find that the determinant of
can be computed using a familiar matrix representation:
(33)
Inverses and adjugates
The inverse of a second-order tensor
is the tensor that satisfies
(34)
For the inverse of to exist, its determinant
. Taking the transpose of (34), we find that the inverse of the transpose of
is the transpose of the inverse. The adjugate
satisfies
(35)
for any two vectors and
. If
is invertible, then (35) yields a relationship between
and
:
(36)
Eigenvalues and eigenvectors
The eigenvalues (or characteristic values, or principal values) of a second-order tensor are defined as the roots
of the characteristic equation
(37)
The three roots of this equation are denoted by ,
, and
. On expanding the characteristic equation (37), we find that
(38)
where
(39)
The corresponding eigenvectors (or characteristic directions, or principal directions) of are the vectors
that satisfy
(40)
A second-order tensor has three eigenvectors, one associated with each eigenvalue. To determine these eigenvectors, we express (40), with the help of (20), in matrix-vector form and then use standard solution techniques from linear algebra. Note that the eigenvectors are unique up to a multiplicative constant.
Proper-orthogonal tensors
A second-order tensor is said to be orthogonal if
. That is, the transpose of an orthogonal tensor is its inverse. It also follows that
. An orthogonal tensor
has the unique property that
for any vector
, and so it preserves the length of the vector that it transforms. A tensor
is proper-orthogonal if it is orthogonal and its determinant
specifically. Thus, proper-orthogonal second-order tensors are a subclass of the second-order orthogonal tensors.
Positive-definite tensors
A second-order tensor is said to be positive-definite if
for any vector
and
if, and only if,
. A consequence of this definition is that a skew-symmetric second-order tensor can never be positive-definite. If
is positive-definite, then all three of its eigenvalues are positive and, furthermore, the tensor has the representation
(41)
where and
are the eigenvalues and eigenvectors of
, respectively. This representation is often known as the spectral decomposition.
Third-order tensors
A third-order tensor transforms vectors into second-order tensors and may transform second-order tensors into vectors. With respect to a right-handed orthonormal basis , any third-order tensor
can be represented as
(42)
and we define the following two tensor products:
(43)
Note the presence of the brackets in (43)1. The main example of a third-order tensor we use throughout this site is the alternator
:
(44)
This tensor has some useful features. First, if is a symmetric second-order tensor, then
. Second, for any vector
,
(45)
which is a skew-symmetric second-order tensor. Thus, can be used to transform a vector into a second-order skew-symmetric tensor and transform the skew-symmetric part of a second-order tensor into a vector.
Axial vectors
The fact that the third-order alternating tensor acts on a vector to produce a skew-symmetric second-order tensor enables us to define a skew-symmetric tensor
for every vector
, and vice versa:
(46)
The vector is known as the axial vector of
. Notice that if
has the representation
(47)
then, with the help of (43)1, its axial vector
(48)
We also note the important result
(49)
for any other vector . This identity allows us to replace cross products with tensor products, and vice versa. We often express the relationship between a skew-symmetric tensor
and its axial vector
without explicit mention of the alternator
:
(50)
Differentiation of tensors
One often encounters derivatives of tensors. Suppose a second-order tensor has the representation (13), where the tensor components
and the basis vectors
are functions of time. The time derivative of
is defined as
(51)
Notice that we differentiate both the components and the basis vectors. We can also define a chain rule and product rules. If the tensors and
and the vector
, then
(52)
Now, consider a function . The derivative of this function with respect to
is defined to be the second-order tensor
(53)
In addition, if the basis vectors are constant, then
(54)
References
- Casey, J., A treatment of rigid body dynamics, ASME Journal of Applied Mechanics 50(4a) 905–907 and 51 227 (1983).
- Casey, J., On the advantages of a geometrical viewpoint in the derivation of Lagrange’s equations for a rigid continuum, Zeitschrift für angewandte Mathematik und Physik 46 S805–S847 (1995).
- Chadwick, P., Continuum Mechanics: Concise Theory and Problems, Dover Publications, New York (1999). Reprint of the George Allen & Unwin Ltd., London, 1976 edition.
- Gurtin, M. E., An Introduction to Continuum Mechanics, Academic Press, New York (1981).