In physics, a covariant transformation is a rule that specifies how certain entities, such as vectors or tensors, change under a change of basis.[1] The transformation that describes the new basis vectors as a linear combination of the old basis vectors is defined as a covariant transformation. Conventionally, indices identifying the basis vectors are placed as lower indices and so are all entities that transform in the same way. The inverse of a covariant transformation is a contravariant transformation. Whenever a vector should be invariant under a change of basis, that is to say it should represent the same geometrical or physical object having the same magnitude and direction as before, its components must transform according to the contravariant rule. Conventionally, indices identifying the components of a vector are placed as upper indices and so are all indices of entities that transform in the same way. The sum over pairwise matching indices of a product with the same lower and upper indices is invariant under a transformation.
A vector itself is a geometrical quantity, in principle, independent (invariant) of the chosen basis. A vector v is given, say, in components vi on a chosen basis ei. On another basis, say e′j, the same vector v has different components v′j andAs a vector, v should be invariant to the chosen coordinate system and independent of any chosen basis, i.e. its "real world" direction and magnitude should appear the same regardless of the basis vectors. If we perform a change of basis by transforming the vectors ei into the basis vectors e′j, we must also ensure that the components vi transform into the new components v′j to compensate.
The needed transformation of v is called the contravariant transformation rule.
In the shown example, a vector is described by two different coordinate systems: a rectangular coordinate system (the black grid), and a radial coordinate system (the red grid). Basis vectors have been chosen for both coordinate systems: ex and ey for the rectangular coordinate system, and er and eφ for the radial coordinate system. The radial basis vectors er and eφ appear rotated anticlockwise with respect to the rectangular basis vectors ex and ey. The covariant transformation, performed to the basis vectors, is thus an anticlockwise rotation, rotating from the first basis vectors to the second basis vectors.
The coordinates of v must be transformed into the new coordinate system, but the vector v itself, as a mathematical object, remains independent of the basis chosen, appearing to point in the same direction and with the same magnitude, invariant to the change of coordinates. The contravariant transformation ensures this, by compensating for the rotation between the different bases. If we view v from the context of the radial coordinate system, it appears to be rotated more clockwise from the basis vectors er and eφ. compared to how it appeared relative to the rectangular basis vectors ex and ey. Thus, the needed contravariant transformation to v in this example is a clockwise rotation.
The explicit form of a covariant transformation is best introduced with the transformation properties of the derivative of a function. Consider a scalar function f (like the temperature at a location in a space) defined on a set of points p, identifiable in a given coordinate system
xi, i=0,1,...
{x'}j,j=0,1,...
{x}i
xi\left({x'}j\right),j=0,1,...
\partialf | |
\partial{x |
i}=
\partialf | |
\partial{x' |
j}
\partial{x' | |
j}{\partial |
{x}i}
This is the explicit form of the covariant transformation rule. The notation of a normal derivative with respect to the coordinates sometimes uses a comma, as follows
f,i \stackrel{def
A vector can be expressed in terms of basis vectors. For a certain coordinate system, we can choose the vectors tangent to the coordinate grid. This basis is called the coordinate basis.
To illustrate the transformation properties, consider again the set of points p, identifiable in a given coordinate system
xi
i=0,1,...
f \left(x0,x1,...\right)
dc/dλ
v[f] \stackrel{def
The parallel between the tangent vector and the operator can also be worked out in coordinates
v[f]=
dxi | |
dλ |
\partialf | |
\partialxi |
or in terms of operators
\partial/\partialxi
v=
dxi | |
dλ |
\partial | |
\partialxi |
=
dxi | |
dλ |
ei
ei=\partial/\partialxi
If we adopt a new coordinates system
{x'}i, i=0,1,...
{xi}
xi\left({x'}j\right),j=0,1,...
e'i={\partial}/{\partial{x'}i}
ei
e'i=
\partial | |
\partial{x' |
i}=
\partialxj | |
\partial{x' |
i}
\partial | |
\partialxj |
=
\partialxj | |
\partial{x' |
i}ej
The components of a (tangent) vector transform in a different way, called contravariant transformation. Consider a tangent vector v and call its components
vi
ei
e'i
{v'}i
v=viei={v'}ie'i
vi=
dxi | |
dλ |
and {v'}i=
d{x' | |
i}{dλ} |
If we express the new components in terms of the old ones, then
{v'}i=
d{x' | |
i}{dλ } |
=
\partial{x' | |
i}{\partial |
xj}
dxj | |
dλ |
=
\partial{x' | |
i}{\partial |
xj}{v}j
An example of a contravariant transformation is given by a differential form df. For f as a function of coordinates
xi
dxi
d{x'}i=
\partial{x' | |
i}{\partial |
{x}j}{dx}j
Entities that transform covariantly (like basis vectors) and the ones that transform contravariantly (like components of a vector and differential forms) are "almost the same" and yet they are different. They have "dual" properties.What is behind this, is mathematically known as the dual space that always goes together with a given linear vector space.
Take any vector space T. A function f on T is called linear if, for any vectors v, w and scalar α:
\begin{align} f(v+w)&=f(v)+f(w)\\ f(\alphav)&=\alphaf(v) \end{align}
A simple example is the function which assigns a vector the value of one of its components (called a projection function). It has a vector as argument and assigns a real number, the value of a component.
All such scalar-valued linear functions together form a vector space, called the dual space of T. The sum f+g is again a linear function for linear f and g, and the same holds for scalar multiplication αf.
Given a basis
ei
ei
\omega0
e0
{\omega}0
v=viei
\omega0(v)=\omega0(viei)=vi
0(e | |
\omega | |
i) |
=v0
There are as many dual basis vectors
\omegai
ei
Sometimes an extra notation is introduced where the real value of a linear function σ on a tangent vector u is given as
\sigma[u]:=\langle\sigma,u\rangle
\langle\sigma,u\rangle
A tensor of type (r, s) may be defined as a real-valued multilinear function of r dual vectors and s vectors. Since vectors and dual vectors may be defined without dependence on a coordinate system, a tensor defined in this way is independent of the choice of a coordinate system.
The notation of a tensor is
\begin{align} &T\left(\sigma,\ldots,\rho,u,\ldots,v\right)\\ \equiv{}&{T\sigma
u,v
Because a tensor depends linearly on its arguments, it is completely determined if one knows the values on a basis
\omegai\ldots\omegaj
ek\ldotsel
T(\omegai,\ldots,\omegaj,ek\ldotsel)= {Ti\ldots
{Ti\ldots
If we choose another basis (which are a linear combination of the original basis), we can use the linear properties of the tensor and we will find that the tensor components in the upper indices transform as dual vectors (so contravariant), whereas the lower indices will transform as the basis of tangent vectors and are thus covariant. For a tensor of rank 2, we can verify that
{A'}i=
\partialxl | |
\partial{x' |
i}
\partialxm | |
\partial{x' |
j}Al
{A'}i=
\partial{x' | |
i}{\partial |
xl}
\partial{x' | |
j}{\partial |
xm}Al
For a mixed co- and contravariant tensor of rank 2
i{} | |
{A'} | |
j= |
\partial{x' | |
i} |
{\partialxl}
\partialxm | |
\partial{x' |
j}
l{} | |
A | |
m |