In mathematics, Ricci calculus constitutes the rules of index notation and manipulation for tensors and tensor fields on a differentiable manifold, with or without a metric tensor or connection.
A component of a tensor is a real number that is used as a coefficient of a basis element for the tensor space. The tensor is the sum of its components multiplied by their corresponding basis elements. Tensors and tensor fields can be expressed in terms of their components, and operations on tensors and tensor fields can be expressed in terms of operations on their components. The description of tensor fields and operations on them in terms of their components is the focus of the Ricci calculus. This notation allows an efficient expression of such tensor fields and operations. While much of the notation may be applied with any tensors, operations relating to a differential structure are only applicable to tensor fields. Where needed, the notation extends to components of non-tensors, particularly multidimensional arrays.
A tensor may be expressed as a linear sum of the tensor product of vector and covector basis elements. The resulting tensor components are labelled by indices of the basis. Each index has one possible value per dimension of the underlying vector space. The number of indices equals the degree (or order) of the tensor.
For compactness and convenience, the Ricci calculus incorporates Einstein notation, which implies summation over indices repeated within a term and universal quantification over free indices. Expressions in the notation of the Ricci calculus may generally be interpreted as a set of simultaneous equations relating the components as functions over a manifold, usually more specifically as functions of the coordinates on the manifold. This allows intuitive manipulation of expressions with familiarity of only a limited set of rules.
Working with a main proponent of the exterior calculus Élie Cartan, the influential geometer Shiing-Shen Chern summarizes the role of tensor calculus:
In our subject of differential geometry, where you talk about manifolds, one difficulty is that the geometry is described by coordinates, but the coordinates do not have meaning. They are allowed to undergo transformation. And in order to handle this kind of situation, an important tool is the so-called tensor analysis, or Ricci calculus, which was new to mathematicians. In mathematics you have a function, you write down the function, you calculate, or you add, or you multiply, or you can differentiate. You have something very concrete. In geometry the geometric situation is described by numbers, but you can change your numbers arbitrarily. So to handle this, you need the Ricci calculus.
Some sources use 4 instead of 0 as the index value corresponding to time; in this article, 0 is used. Otherwise, in general mathematical contexts, any symbols can be used for the indices, generally running over all dimensions of the vector space.
For example, in 3-D Euclidean space and using Cartesian coordinates; the coordinate vector shows a direct correspondence between the subscripts 1, 2, 3 and the labels , , . In the expression , is interpreted as an index ranging over the values 1, 2, 3, while the , , subscripts are only labels, not variables. In the context of spacetime, the index value 0 conventionally corresponds to the label .
This is not to be confused with van der Waerden notation for , which uses hats and overdots on indices to reflect the chirality of a spinor.
In the special case that the metric tensor is everywhere equal to the identity matrix, it is possible to drop the distinction between upper and lower indices, and then all indices could be written in the lower position. Coordinate formulae in linear algebra such as for the product of matrices may be examples of this. But in general, the distinction between upper and lower indices should be maintained.
Ordering of indices is significant, even when of differing variance. However, when it is understood that no indices will be raised or lowered while retaining the base symbol, covariant indices are sometimes placed below contravariant indices for notational convenience (e.g. with the generalized Kronecker delta).
The number of indices of a tensor, regardless of variance, is called the degree of the tensor (alternatively, its valence, order or rank, although rank is ambiguous). Thus, a tensor of type has degree .
The operation implied by such a summation is called tensor contraction:
This summation may occur more than once within a term with a distinct symbol per pair of indices, for example:
Other combinations of repeated indices within a term are considered to be ill-formed, such as
| (both occurrences of are lower; would be fine) |
| ( occurs twice as a lower index; or would be fine). |
A_
A_{\alpha \beta \gamma \cdots} B^{|\alpha\beta\gamma \cdots} =
\sum_{\alpha < \beta < \gamma} A_{\alpha \beta \gamma \cdots} B^{\alpha\beta\gamma \cdots}
means a restricted sum over index values, where each index is constrained to being strictly less than the next. More than one group can be summed in this way, for example:
&A_{ |
When using multi-index notation, an underarrow is placed underneath the block of indices:
A_{\underset{\rightharpoondown}{P}}{}^{\underset{\rightharpoondown}{Q}} B^P{}_{Q\underset{\rightharpoondown}{R}} C^R =
\sum_\underset{\rightharpoondown}{P} \sum_\underset{\rightharpoondown}{Q} \sum_\underset{\rightharpoondown}{R} A_{P}{}^{Q} B^P{}_{QR} C^R
where
\underset{\rightharpoondown}{P} = |\alpha \beta\gamma|\,,\quad
\underset{\rightharpoondown}{Q} = |\delta\epsilon\cdots\lambda|\,,\quad
\underset{\rightharpoondown}{R} = |\mu \nu \cdots\zeta|
The base symbol in many cases is retained (e.g. using where appears here), and when there is no ambiguity, repositioning an index may be taken to imply this operation.
The Kronecker delta is used, see also below.
for all . Consequently, there are facets of the notation that are useful in checking that an equation makes sense (an analogous procedure to dimensional analysis).
For instance, if
A^0 B_1{}^0 C_{00} + A^0 B_1{}^1 C_{10} + A^0 B_1{}^2 C_{20} + A^0 B_1{}^3 C_{30} + D^0{}_1{} E_0 &= T^0{}_1{}_0 \\
A^1 B_0{}^0 C_{00} + A^1 B_0{}^1 C_{10} + A^1 B_0{}^2 C_{20} + A^1 B_0{}^3 C_{30} + D^1{}_0{} E_0 &= T^1{}_0{}_0 \\
A^1 B_2{}^0 C_{02} + A^1 B_2{}^1 C_{12} + A^1 B_2{}^2 C_{22} + A^1 B_2{}^3 C_{32} + D^1{}_2{} E_2 &= T^1{}_2{}_2.
\end{align}
This illustrates the compactness and efficiency of using index notation: many equations which all share a similar structure can be collected into one simple tensor equation.
In the first replacement, replaced and replaced everywhere, so the expression still has the same meaning. In the second, did not fully replace , and did not fully replace (incidentally, the contraction on the index became a tensor product), which is entirely inconsistent for reasons shown next.
as for an erroneous expression:
In other words, non-repeated indices must be of the same type in every term of the equation. In the above identity, line up throughout and occurs twice in one term due to a contraction (once as an upper index and once as a lower index), and thus it is a valid expression. In the invalid expression, while lines up, and do not, and appears twice in one term (contraction) and once in another term, which is inconsistent.
If the brackets enclose covariant indices – the rule applies only to all covariant indices enclosed in the brackets, not to any contravariant indices which happen to be placed intermediately between the brackets.
Similarly if brackets enclose contravariant indices – the rule applies only to all enclosed contravariant indices, not to intermediately placed covariant indices.
A_{(\alpha_1\alpha_2\cdots\alpha_p)\alpha_{p + 1}\cdots\alpha_q} =
\dfrac{1}{p!} \sum_{\sigma} A_{\alpha_{\sigma(1)}\cdots\alpha_{\sigma(p)}\alpha_{p + 1}\cdots\alpha_{q}} \,.
For example, two symmetrizing indices mean there are two indices to permute and sum over:
A_{(\alpha\beta\gamma)\delta\cdots} =
\dfrac{1}{3!} \left(A_{\alpha\beta\gamma\delta\cdots}
+ A_{\gamma\alpha\beta\delta\cdots}
+ A_{\beta\gamma\alpha\delta\cdots}
+ A_{\alpha\gamma\beta\delta\cdots}
+ A_{\gamma\beta\alpha\delta\cdots}
+ A_{\beta\alpha\gamma\delta\cdots}
\right)
The symmetrization is distributive over addition;
Indices are not part of the symmetrization when they are:
Here the and indices are symmetrized, is not.
& A_{[\alpha_1\cdots\alpha_p]\alpha_{p+1}\cdots\alpha_q} \\[3pt]
={} & \dfrac{1}{p!} \sum_{\sigma}\sgn(\sigma) A_{\alpha_{\sigma(1)}\cdots\alpha_{\sigma(p)}\alpha_{p+1}\cdots\alpha_{q}} \\
={} & \delta_{\alpha_1 \cdots \alpha_p}^{\beta_1 \dots \beta_p} A_{\beta_1 \cdots \beta_p\alpha_{p+1}\cdots\alpha_q} \\
\end{align}
where is the generalized Kronecker delta of degree , with scaling as defined below.
For example, two antisymmetrizing indices imply:
A_{[\alpha\beta\gamma]\delta\cdots} =
\dfrac{1}{3!} \left(A_{\alpha\beta\gamma\delta\cdots}
+ A_{\gamma\alpha\beta\delta\cdots}
+ A_{\beta\gamma\alpha\delta\cdots}
- A_{\alpha\gamma\beta\delta\cdots}
- A_{\gamma\beta\alpha\delta\cdots}
- A_{\beta\alpha\gamma\delta\cdots}
\right)
as for a more specific example, if represents the electromagnetic tensor, then the equation
0 = F_{[\alpha\beta, \gamma]} = \dfrac{1}{3!} \left(
F_{\alpha\beta,\gamma}
+ F_{\gamma\alpha,\beta}
+ F_{\beta\gamma,\alpha}
- F_{\beta\alpha,\gamma}
- F_{\alpha\gamma,\beta}
- F_{\gamma\beta,\alpha}
\right) \,
represents Gauss's law for magnetism and Faraday's law of induction.
As before, the antisymmetrization is distributive over addition;
A_{[\alpha} \left(B_{\beta]\gamma\cdots} + C_{\beta]\gamma\cdots} \right) =
A_{[\alpha}B_{\beta]\gamma\cdots} + A_{[\alpha}C_{\beta]\gamma\cdots}
As with symmetrization, indices are not antisymmetrized when they are:
A_{[\alpha}B^{\beta}{}_{\gamma]} =
\dfrac{1}{2!} \left(A_{\alpha}B^{\beta}{}_{\gamma} - A_{\gamma}B^{\beta}{}_{\alpha} \right)
A_{[\alpha}B_Here the and indices are antisymmetrized, is not.
To indicate partial differentiation of the components of a tensor field with respect to a coordinate variable , a comma is placed before an appended lower index of the coordinate variable.
This may be repeated (without adding further commas):
A_{\alpha_1\alpha_2\cdots\alpha_p\,,\,\alpha_{p+1}\cdots\alpha_q} =
\dfrac{\partial}{\partial x^{\alpha_q}}\cdots\dfrac{\partial}{\partial x^{\alpha_{p+2}}}\dfrac{\partial}{\partial x^{\alpha_{p+1}}} A_{\alpha_1\alpha_2\cdots\alpha_p}.
These components do not transform covariantly, unless the expression being differentiated is a scalar. This derivative is characterized by the product rule and the derivatives of the coordinates
The covariant derivative of a scalar function, a contravariant vector and a covariant vector are:
T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s ; \gamma}
& \\
= T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s , \gamma}
&+ \, \Gamma^{\alpha_1}{}_{\delta \gamma} T^{\delta \alpha_2 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s} + \cdots + \Gamma^{\alpha_r}{}_{\delta \gamma} T^{\alpha_1 \cdots \alpha_{r-1} \delta}{}_{\beta_1 \cdots \beta_s} \\
&- \, \Gamma^\delta{}_{\beta_1 \gamma} T^{\alpha_1 \cdots \alpha_r}{}_{\delta \beta_2 \cdots \beta_s} - \cdots - \Gamma^\delta{}_{\beta_s \gamma} T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_{s-1} \delta}\,.
\end{align}
An alternative notation for the covariant derivative of any tensor is the subscripted nabla symbol . For the case of a vector field :
The covariant formulation of the directional derivative of any tensor field along a vector may be expressed as its contraction with the covariant derivative, e.g.:
The components of this derivative of a tensor field transform covariantly, and hence form another tensor field, despite subexpressions (the partial derivative and the connection coefficients) separately not transforming covariantly.
This derivative is characterized by the product rule:
A connection is a metric connection when the covariant derivative of the metric tensor vanishes:
An affine connection that is also a metric connection is called a Riemannian connection. A Riemannian connection that is torsion-free (i.e., for which the torsion tensor vanishes: ) is a Levi-Civita connection.
The for a Levi-Civita connection in a coordinate basis are called Christoffel symbols of the second kind.
This derivative is not defined on any tensor field with contravariant indices or that is not totally antisymmetric. It is characterized by a graded product rule.
(\mathcal{L}_X T)^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s}
& \\
= X^\gamma T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s , \gamma}
& - \, X^{\alpha_1}{}_{, \gamma} T^{\gamma \alpha_2 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s} - \cdots - X^{\alpha_r}{}_{, \gamma} T^{\alpha_1 \cdots \alpha_{r-1} \gamma}{}_{\beta_1 \cdots \beta_s} \\
& + \, X^{\gamma}{}_{, \beta_1} T^{\alpha_1 \cdots \alpha_r}{}_{\gamma \beta_2 \cdots \beta_s} + \cdots + X^{\gamma}{}_{, \beta_s} T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_{s-1} \gamma} \,.
\end{align}
This derivative is characterized by the product rule and the fact that the Lie derivative of a contravariant vector field along itself is zero:
\delta^{\alpha}_{\beta} \, A^{\beta} &= A^{\alpha} \\
\delta^{\mu}_{\nu} \, B_{\mu} &= B_{\nu} .
\end{align}
The components are the same in any basis and form an invariant tensor of type , i.e. the identity of the tangent bundle over the identity mapping of the base manifold, and so its trace is an invariant. Its trace is the dimensionality of the space; for example, in four-dimensional spacetime,
The Kronecker delta is one of the family of generalized Kronecker deltas. The generalized Kronecker delta of degree may be defined in terms of the Kronecker delta by (a common definition includes an additional multiplier of on the right):
For a Levi-Civita connection this tensor is defined to be zero, which for a coordinate basis gives the equations
- \Gamma^\rho{}_{\mu\sigma,\nu}
+ \Gamma^\rho{}_{\mu\lambda}\Gamma^\lambda{}_{\nu\sigma}
- \Gamma^\rho{}_{\nu\lambda}\Gamma^\lambda{}_{\mu\sigma} \,,
then it is the commutator of the covariant derivative with itself:
This can be generalized to get the commutator for two covariant derivatives of an arbitrary tensor as follows:
T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s ; \gamma \delta}&
- T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s ; \delta \gamma} \\
&\!\!\!\!\!\!\!\!\!\!= - R^{\alpha_1}{}_{\rho \gamma \delta} T^{\rho \alpha_2 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_s}
- \cdots
- R^{\alpha_r}{}_{\rho \gamma \delta} T^{\alpha_1 \cdots \alpha_{r-1} \rho}{}_{\beta_1 \cdots \beta_s} \\
&+ R^\sigma{}_{\beta_1 \gamma \delta} T^{\alpha_1 \cdots \alpha_r}{}_{\sigma \beta_2 \cdots \beta_s}
+ \cdots
+ R^\sigma{}_{\beta_s \gamma \delta} T^{\alpha_1 \cdots \alpha_r}{}_{\beta_1 \cdots \beta_{s-1} \sigma} \,
\end{align}
which are often referred to as the Ricci identities.
The inverse matrix of the metric tensor is another important tensor, used for raising indices:
|
|