# Kronecker Delta: 4 Important Rules and Scalar Product in Index Notation

## Table of contents

- Definition and Examples Mathematical definition of the Kronecker delta with some examples.
- Einstein's Summation Convention Here you will learn about the Einstein summation convention, where you can omit the sum signs and get formal commutativity and compactness.
- 4 Rules for Kronecker delta Here you will learn four important calculation rules when using the Kronecker delta.
- The 3 most common mistakes you should avoid making Here you will learn the common mistakes that are made when calculating with the Kronecker delta.
- Scalar product with Kronecker delta Here you will learn how to write the scalar product in index notation using the Kronecker delta.

## Video - Kronecker delta explained simply in 11 minutes!

Subscribe on YouTubeIt is impossible to imagine theoretical physics without the Kronecker delta. You will encounter this relatively simple, yet powerful tensor practically in all fields of theoretical physics. For example, it is used to

write long expressions more compactly.

simplify complicated expressions.

In combination with the *Levi-Civita tensor*, the two tensors are very powerful! That's why it's worth understanding how the Kronecker delta works.

## Definition and Examples

Kronecker delta \(\delta_{ \class{blue}{i} \class{red}{j} }\) - is a small greek letter delta, which yields either 1 or 0, depending on which values its two indices \(\class{blue}{i} \) and \(\class{red}{j}\) take on. The maximal value of an index corresponds to the considered dimension, so in three-dimensional space \(\class{blue}{i} \) and \(\class{red}{j}\) run from 1 to 3.

Kronecker delta is equal to 1 if \( \class{blue}{i} \) and \(\class{red}{j}\) are *equal*. And Kronecker delta is equal to 0 if \(\class{blue}{i} \) and \(\class{red}{j}\) are *unequal*.

**Definition: Kronecker delta**

0, &\mbox{} \class{blue}{i} \neq \class{red}{j} \end{cases} \end{align} $$

## Einstein's Summation Convention

In order to represent an expression like

**Summation with sum sign over an index in the product of two Kronecker deltas**

\end{align} $$

or an expression like

**Summation with sum sign over an index in the product of two vector components**

compactly, we agree on the following rule:

Another advantage of the sum convention (in addition to compactness) is formal *commutativity*. For example, you may write down the expression \( \varepsilon_{ijk} \, \boldsymbol{\hat{e}}_{i} \, s_j \, \delta_{km} \) in a different order as you wish, for example like this: \( \varepsilon_{ijk} \, \delta_{km} \, \, s_j \, \boldsymbol{\hat{e}}_{i} \). This might help you see what can be shortened or simplified further.

**But be careful!** There are exceptions, for example with the differential operator \( \partial_j \), which acts on a successor.

**Derivative operator does not commutate in index notation**

\end{align} $$

You can' t move something before the derivative that is supposed to be differentiated. So you should be careful with operators in index notation.

## 4 Rules for Kronecker delta

Let's look at four useful calculation rules with Kronecker delta that you can use whenever summing over double indices.

**Why is that?**: According to the definition, if the indices \(i\) and \(j\) are equal, then \( \delta_{ij} \) is equal to 1. But then, also \(\delta_{ji} \) is equal to 1. And if the indices are unequal, you have: \( \delta_{ij} \) is equal to zero. And \(\delta_{ji} \) is equal to zero. So as you can see: Kronecker delta is *symmetric*!

**Why is that?**: Let's consider for example the case where the indices \( i \) and \( j \) are equal: \( i = j \) and the indices \( j \) and \( k \) are not equal: \( j \neq k \). Then it follows that \( i \) and \( k \) must also be unequal: \( i \neq k \). So \(\delta_{jk}\) is zero and therefore the whole term on the left hand side is zero: \( \delta_{ij} \, \delta_{jk} ~=~ 0 \). \(\delta_{ik}\) on the left-hand side is also zero, because \(i\) and \(k\) are different. The equation is fulfilled. You can proof all other possible cases in the same way. So instead of writing two deltas you can just write \( \delta_{ik} \). We say: The summation index \(j\) is *contracted*.

**Why is that?** This rule is basically another case of index contraction. This rules tells you that you can also contract summation indices that don't have to be carried by a Kronecker delta.

**Why is that?**: According to the summation convention 5

, the summation is carried out over \(j\) here. So \(\delta_{jj}\) is equal to \(\delta_{11}\) plus \(\delta_{22}\) plus \(\delta_{33}\) and so on up to \(n\). And each Kronecker delta yields 1, because the index values are equal. So 1 + 1 + 1 and so on, results in \(n\):

**Write out Kronecker delta with two equal indices**

&~=~ 1 ~+~ 1 ~+~...~+~ 1 \\\\

& ~=~ n \end{align} $$

## The 3 most common mistakes you should avoid making

If you use summation convention and the above rules, you must also pay attention to the *correct notation*: Summation is done if an index occurs exactly *twice* on *one* side of the equation.

**Mistake #1**Formula anchor $$ \begin{align} v_{\class{blue}{i}} ~=~ b_{\class{blue}{i}} \, c_{\class{blue}{i}} \end{align} $$**Why?**Because on the right hand side you have double index \(i\), so you sum over \( i \). But on the left hand side there is also the summation index \(i\) - and this of course makes no sense... To correct this expression, you can rename the summation index to \( j \).This expression is not formally wrong but very prone to errors:

**Mistake #2**Formula anchor $$ \begin{align} \delta_{\class{red}{j}\class{green}{k}} \, v_{\class{green}{k}} ~=~ \delta_{\class{violet}{m}\class{green}{k}} \, r_{\class{green}{k}}

\end{align} $$Because the summation index \( k \) appears on

*both*sides of the equation. To correct this, rename one of the summation indices, for example to \(j\).**Mistake #3**Formula anchor $$ \begin{align} \delta_{\class{blue}{i}1} \, \delta_{1\class{green}{k}} ~=~ \delta_{\class{blue}{i}\class{green}{k}}

\end{align} $$**Why?**Because here you try to sum over an ordinary*number*, as if this number would be a summation index. Number "1" occurs twice, but it is*not a variable index*over which you can sum. Therefore you can not use the contraction rule on ordinary numbers.

## Scalar product with Kronecker delta

Consider a three-dimensional vector with the components \(x\), \(y\) and \(z\):

**Column vector with three components**

You can represent this vector \( \boldsymbol{v} \) in an orthonormal basis as follows:

**Vector as linear combination of basis vectors**

\end{align} $$

Here \(\boldsymbol{\hat{e}}_x\), \(\boldsymbol{\hat{e}}_y\) and \(\boldsymbol{\hat{e}}_z\) are three basis vectors which are orthogonal to each other and normalized. In this case they span an orthogonal three-dimensional coordinate system.

The Kronecker delta needs vectors written in **index notation**. Here we do not denote the vector components with different letters \(x,y,z\), but we choose one letter (here the letter \( v \)) and then number the vector components consecutively. The vector components are then called \(v_1\), \(v_2\) and \(v_3\) and the basis expansion looks like this:

**Vector as linear combination of basis vectors in other notation**

&~=~ v_1 \, \boldsymbol{\hat{e}}_1 ~+~ v_2 \, \boldsymbol{\hat{e}}_2 ~+~ v_3 \, \boldsymbol{\hat{e}}_3 \end{align} $$

One of the advantages of index notation is that this way you will never run out of letters for the vector components. Just imagine a fifty-dimensional vector. There aren't even that many letters to give each component \(v_1\), \(v_2\) ... \(v_{50}\) of the vector a unique letter! It gets worse when you want to write out this fifty-dimensional vector as in 21

....

Another advantage of the index notation is that by numbering the vector components in this way, you can use the sum sign to represent 21

more compactly. It becomes even more compact if we omit the big sum sign according to the summation convention. Look how compact the vector \(\boldsymbol{v}\) can be represented in a basis:

**Vector in index notation**

Here, as you know, we sum over index \(j\). Whether you call the index \(j\), \(i\) or \(k\) or any other letter is of course up to you.

Now that you know how a vector is represented in index notation, we can analogously write the **scalar product** \( \boldsymbol{a} ~\cdot~ \boldsymbol{b} \) of two vectors \( \boldsymbol{a} = (a_1, a_2, a_3) \) and \( \boldsymbol{b} = (b_1, b_2, b_3)\) in index notation. For this we use the index representation of a vector 22

**Scalar product of two vectors in index notation**

In index notation, you may sort the factors in 23

as you like. This is the advantage of index notation, where the commutative law applies. Let's take advantage of that and put parentheses around the basis vectors to emphasize their importance in introducing the Kronecker delta:

**Scalar product of two vectors in index notation with factored out basis vectors**

The basis vectors \(\boldsymbol{\hat{e}}_i\) and \(\boldsymbol{\hat{e}}_j\) are *orthonormal* (i.e. orthogonal and normalized). Recall what the property of being orthonormal means for two vectors. Their scalar product \( \boldsymbol{\hat{e}}_{i} ~\cdot~ \boldsymbol{\hat{e}}_{j} \) yields:

**Scalar product of two basis vectors in index notation**

0, &\mbox{} \class{blue}{i} \neq \class{red}{j} \end{cases} \end{align} $$

Doesn't this property look familiar to you? The scalar product of two orthonormal vectors behaves *exactly like Kronecker delta*!

**Definition of Kronecker delta is like scalar product of two basis vectors in index notation**

0, &\mbox{} \class{blue}{i} \neq \class{red}{j} \end{cases} \end{align} $$

Therefore replace the scalar product of two basis vectors with a Kronecker delta:

**Scalar product of two orthonormal vectors expressed with Kronecker delta**

Thus we can write the scalar product 27

using Kronecker delta:

**Scalar product with Kronecker delta**

\end{align} $$

If you remember rule #3, you can contract the index \(j\) if you want:

**Scalar product with eliminated Kronecker delta**

And you get exactly the definition of the scalar product, where the vector components are summed component-wise:

**Scalar product of two vectors written out in index notation**

\end{align} $$