# 1. Einstein summation convention

**Einstein notation**, or **Einstein summation convention**, is simply a reduced form of well-known summation notation introduced by Albert Einstein in 1916. For example, given two vectors , we write the inner product as in new notation .** **At the first glance there is nothing special as just omit the summation notation (This is exactly what I feel when I first saw the notation). But I will show you this reduction brings much more than convenience. Moreover, it indicates the object which the component belongs to. Specifically speaking, it distinguishes the type of the tensor.

Before given the formal statement of the convention, let’s start with a few examples. We will denote vectors of dimension by lower letters and matrices of proper dimensions by capital letters .

- inner product:
- bilinear form:
- linear transformation:
- matrix multiplication:
- trace of matrix:

Several remarks should be noted here. First, we see summation is taken **only** for those indices that repeat. And the repeated indices **always** occur in pairs, one in upper and the other in lower position. Once the summation is taken, all possible values of the repeated indices should be contained. So it makes nonsense of expressing .

Second, **all** indices, including repeated and non-repeated indices, are compatible. Note that repeated indices disappear in the result (left hand side of the identity). We call this sort of indices dummy since they represent nothing in the result, which implies we can replace dummy index by any other allowable (this shan’t conflict with the existed indices) letter. That is, is equivalent to , but not to . As for non-repeated indices, they appear at the same time on both sides of the identity, and at the same position, both in upper or both in lower positions. Note that both superscript and subscript are indices rather than powers. Say, we always use to denote the second component of vector rather than squared.

Third, it is readily to verify (leave to readers) that each component, as a scalar, satisfies all arithmetic laws for a field, ie.

With Einstein notation, we can pay more attention on algebraic computing than checking consistency and then deciding appropriate operations between terms because everything works well all the way that is needless to care. We can sometimes surprisingly find some interesting identities which doesn’t seem obvious using notation of vector and matrix algebra. For instance, a series of expressions naturally equates from the above identities, where means inner product with respect to and means inner product of matrices. Till now, we are able to summarize and formally state the following.

**Einstein summation convention**: *In an expression,* s*ummation is automatically taken over all the values of repeated index which occurs in pairs, once in upper and once in lower position.*

Readers may ask why we use both superscript and subscript to represent vectors (there indeed some authors don’t require this). Roughly speaking, a single superscript represents some component of column vector, while a single subscript represents some component of row vector. Column vector space and row vector space are dual to each other in finite dimensional case. Essentially, we adopt superscripts to denote contravariant components and subscripts to denote covariant components. Contravariance and covariance are also a pair of dual concepts implying different transformation laws of tensors when change of basis of base vector space. There is also mix-variance tensor e.g. linear transformation. But I have to put off talking about details on this topic since it worth a whole post.

To tell the truth, I didn’t find it useful in that it is really too simple to give a notice. This opinion gradually turned when I found many applications in matrix calculus for it. The derivative with respect to a vector or a matrix is not so simple as it with respect to scalar because of noncommutativity of matrix. The multiplications of matrices with different orders usually give different results and even different structures. However, with the aid of index notation, the order doesn’t matter and the derivative is just like the original one we familiar with. Then I don’t need to learn and remember those strange rules in matrix calculus. In some sense it is quite enjoyable to solve a “hard” question just like play children’s stuff.

Philosophically, Nothing comes nowhere. Einstein notation is not a simply an abbreviation of summation. I think it implies the essence of linearity. When linear operation is applied on tensor, which is multi-linear object, the linearity suggests we just need to consider each term, or equivalently, the general term. And the sum preserves automatically. The representation focus on micro view of a tensor, indicating explicitly the components and transformation law of tensor. This view regards tensor as nothing but an array of numbers, which is too tight to show off talents of a genius tensor. I personally emphasize the geometric picture of all mathematical concepts and theorems, which provides with some kind of intuition and imagination. Tensor, instead, can also be a linear operator that is coordinate-free. Then we can talk about the domain and range of a tensor, its inverse, adjoint, spectrum etc, in which cases all indices are meaningless. Although not offering geometric picture, Einstein notation itself has enough power surprising everyone from algebraic perspective once introducing only one naive symbol, Kronecker delta.

# 2. Kronecker delta & Levi-Civita symbol

We introduce two symbols now just … for fun. Wait a second, “Ci” in “Civita” is pronounced as “tree”. **Kronecker delta symbol ** is an indicating function of identification of two indices.

where is Iverson bracket giving 1 if holds and 0 otherwise. Kronecker delta looks like identity matrix and plays role of replacing index. For example, leaving invariant and just replacing by . And not only replaces the index, but also pulls down superscript, which can be seen as transpose of the vector. Note that , where is the dimension of vector space.

**Levi-Civita symbol** is defined as the sign of permutation , equivalently, where is the parity of , the number of inversions in . The symbol gives 0 if any two indices of are the same.

Readers have to bear in mind that is not tensor because of its different transformation law, which we call it pseudo-tensor. The interesting thing with Levi-Civita symbol is to compute , where is by matrix. Note for all terms contained in summation, only those terms having components of taken from different rows and columns as factors don’t vanish. Multiplied by sign of permutation , we find that . Surprising, right? As for a direct application in vector algebra, we have , considering the determinant rule for computing cross product.

Another interesting thing also comes from computation. Assume by matrix , where are permutations of order , compute the determinant . According to the above,

Note that and functions as only replace index. This leads us,

By the definition of Levi-Civita symbol, it’s not hard to obtain

Intuitively, gives the sign of permutation . Also, it’s readily to check whenever or for some . The permutation is so common worthy a new symbol, called **generalized Kronecker delt****a**, defined as

Note that integer doesn’t have to be . When , we have . When , the trick is to add dummy indices and consider . By definition, since the last indices are the same, we need only to consider the permutation of the rest indices. With the same reason, summation is automatically taken and produces copies of permutations of the rest indices. Therefore,

Particularly, we have . Let’s see what the role does generalized Kronecker delta play. It doesn’t simply replace the index any more, otherwise , which is obviously wrong. Let , where . Note that alternates sign when interchange any two indices from , actually anti-symmetrizes part or all of components up to a factorial factor!

I think it’s right time to end this post. These notations and symbols are quite simple and common in differential geometry and modern theoretical physics, and also inevitable step to further study. This is my first post, any comments, corrections or suggestions are greatly welcomed.