In mathematics and mathematical physics, **raising and lowering indices** are operations on tensors which change their type. Raising and lowering indices are a form of index manipulation in tensor expressions.

Mathematically vectors are elements of a vector space over a field , and for use in physics is usually defined with or . Concretely, if the dimension of is finite, then, after making a choice of basis, we can view such vector spaces as or .

The dual space is the space of linear functionals mapping . Concretely, in matrix notation these can be thought of as row vectors, which give a number when applied to column vectors. We denote this by , so that is a linear map .

Then under a choice of basis , we can view vectors as an vector with components (vectors are taken by convention to have indices up). This picks out a choice of basis for , defined by the set of relations .

For applications, raising and lowering is done using a structure known as the (pseudo-)metric tensor (the 'pseudo-' refers to the fact we allow the metric to be indefinite). Formally, this is a non-degenerate, symmetric bilinear form

In this basis, it has components , and can be viewed as a symmetric matrix in with these components. The inverse metric exists due to non-degeneracy and is denoted , and as a matrix is the inverse to .

Raising and lowering is then done in coordinates. Given a vector with components , we can contract with the metric to obtain a **covector**:

and this is what we mean by lowering the index. Conversely, contracting a covector with the inverse metric gives a vector:

This process is called raising the index.

Raising and then lowering the same index (or conversely) are inverse operations, which is reflected in the metric and inverse metric tensors being inverse to each other (as is suggested by the terminology):

where is the Kronecker delta or identity matrix.

Finite-dimensional real vector spaces with (pseudo-)metrics are classified up to signature, a coordinate-free property which is well-defined by Sylvester's law of inertia. Possible metrics on real space are indexed by signature . This is a metric associated to dimensional real space. The metric has signature if there exists a basis (referred to as an orthonormal basis) such that in this basis, the metric takes the form with positive ones and negative ones.

The concrete space with elements which are -vectors and this concrete realization of the metric is denoted , where the 2-tuple is meant to make it clear that the underlying vector space of is : equipping this vector space with the metric is what turns the space into .

Examples:

- is a model for 3-dimensional space. The metric is equivalent to the standard dot product.
- , equivalent to dimensional real space as an inner product space with . In Euclidean space, raising and lowering is not necessary due to vectors and covector components being the same.
- is Minkowski space (or rather, Minkowski space in a choice of orthonormal basis), a model for spacetime with weak curvature. It is common convention to use greek indices when writing expressions involving tensors in Minkowski space, while Latin indices are reserved for Euclidean space.

Well-formulated expressions are constrained by the rules of Einstein summation: any index may appear at most once and furthermore a raised index must contract with a lowered index. With these rules we can immediately see that an expression such as

is well formulated while

is not.

The covariant 4-position is given by

with components:

(where x,y,z are the usual Cartesian coordinates) and the Minkowski metric tensor with metric signature (− + + +) is defined as

in components:

To raise the index, multiply by the tensor and contract:

then for *λ* = 0:

and for *λ* = *j* = 1, 2, 3:

So the index-raised contravariant 4-position is:

This operation is equivalent to the matrix multiplication

Given two vectors, and , we can write down their (pseudo-)inner product in two ways:

By lowering indices, we can write this expression as

What is this in matrix notation? The first expression can be written as

while the second is, after lowering the indices of ,

It is instructive to consider what raising and lowering means in the abstract linear algebra setting.

We first fix definitions: is a finite-dimensional vector space over a field . Typically or .

is a non-degenerate bilinear form, that is, is a map which is linear in both arguments, making it a bilinear form.

By being non-degenerate we mean that for each , there is a such that

In concrete applications, is often considered a structure on the vector space, for example an inner product or more generally a metric tensor which is allowed to have indefinite signature, or a symplectic form . Together these cover the cases where is either symmetric or anti-symmetric, but in full generality need not be either of these cases.

There is a partial evaluation map associated to ,

where denotes an argument which is to be evaluated, and denotes an argument whose evaluation is deferred. Then is an element of , which sends .

We made a choice to define this partial evaluation map as being evaluated on the first argument. We could just as well have defined it on the second argument, and non-degeneracy is also independent of argument chosen. Also, when has well defined (anti-)symmetry, evaluating on either argument is equivalent (up to a minus sign for anti-symmetry).

Non-degeneracy shows that the partial evaluation map is injective, or equivalently that the kernel of the map is trivial. In finite dimension, the dual space has equal dimension to , so non-degeneracy is enough to conclude the map is a linear isomorphism. If is a structure on the vector space sometimes call this the canonical isomorphism .

It therefore has an inverse, and this is enough to define an associated bilinear form on the dual:

where the repeated use of is disambiguated by the argument taken. That is, is the inverse map, while is the bilinear form.

Checking these expressions in coordinates makes it evident that this is what raising and lowering indices means abstractly.

We will not develop the abstract formalism for tensors straightaway. Formally, an tensor is an object described via its components, and has components up, components down. A generic tensor is written

We can use the metric tensor to raise and lower tensor indices just as we raised and lowered vector indices and raised covector indices.

- A (0,0) tensor is a number in the field .
- A (1,0) tensor is a vector.
- A (0,1) tensor is a covector.
- A (0,2) tensor is a bilinear form. An example is the metric tensor
- A (1,1) tensor is a linear map. An example is the delta, , which is the identity map, or a Lorentz transformation

For a (0,2) tensor,^{[1]} twice contracting with the inverse metric tensor and contracting in different indices raises each index:

Similarly, twice contracting with the metric tensor and contracting in different indices lowers each index:

Let's apply this to the theory of electromagnetism.

The contravariant electromagnetic tensor in the (+ − − −) signature is given by^{[2]}

In components,

To obtain the covariant tensor F_{αβ}, contract with the inverse metric tensor:

and since *F*^{00} = 0 and *F*^{0i} = − *F ^{i}*

Now for *α* = 0, *β* = *k* = 1, 2, 3:

and by antisymmetry, for *α* = *k* = 1, 2, 3, *β* = 0:

then finally for *α* = *k* = 1, 2, 3, *β* = *l* = 1, 2, 3;

The (covariant) lower indexed tensor is then:

This operation is equivalent to the matrix multiplication

For a tensor of order n, indices are raised by (compatible with above):^{[1]}

and lowered by:

and for a mixed tensor:

We need not raise or lower all indices at once: it is perfectly fine to raise or lower a single index. Lowering an index of an tensor gives a tensor, while raising an index gives a (where have suitable values, for example we cannot lower the index of a tensor.)