Pauli, rotations and quaternions

In Nobody understands quantum mechanics, but spin is fun we met three Pauli matrices \sigma_1,\sigma_2,\sigma_3.. They are kinda OK, but they are not quite fitting our purpose. Therefore I will replace them with another set of matrices, and I will call these matrices s_1,s_2,s_3. They are defined the same way as Pauli matrices, except that s_2 has a different sign:

(1)   \begin{equation*}s_1=\begin{bmatrix}0&1\\1&0\end{bmatrix},\quad s_2=\begin{bmatrix}0&i\\-i&0\end{bmatrix},\quad s_3=\begin{bmatrix}1&0\\0&-1\end{bmatrix}.\end{equation*}

We have

(2)   \begin{equation*} (s_i)^*=s_i,\quad (i=1,2,3),\end{equation*}

(3)   \begin{equation*}(s_1)^2=(s_2)^2=(s_3)^2=I,\end{equation*}

(4)   \begin{equation*}s_1s_2=-is_3,\quad s_2s_3=-is_1,\quad s_3s_1=-is_2.\end{equation*}

The Pauli’s matrices \sigma_i would have in Eq. (4) plus signs on the right. We want minus signs there. Why? Soon we will see why. But, perhaps, I should mention it now, that when Pauli is around many things go not quite right. That is the famous “Pauli effect“. From Wikipedia article Pauli Effect:

The Pauli effect is a term referring to the apparently mysterious, anecdotal failure of technical equipment in the presence of Austrian theoretical physicist Wolfgang Pauli. The term was coined using his name after numerous instances in which demonstrations involving equipment suffered technical problems only when he was present.

One example:

In 1934, Pauli saw a failure of his car during a honeymoon tour with his second wife as proof of a real Pauli effect since it occurred without an obvious external cause.

You can find more examples in the Wikipedia article. The car breakdown during a honeymoon with Pauli’s second wife is somehow related to the change of the sign of the second matrix above. And now it is easy to relate our second set of matrices to quaternions. It was Sir William Hamilton who introduced quaternions in 1843. There were three imaginary units \mathbf{i},\mathbf{j},\mathbf{k} satisfying:

(5)   \begin{equation*}\mathbf{i}^2=\mathbf{j}^2=\mathbf{k}^2=-1,\end{equation*}

(6)   \begin{equation*}\mathbf{i}\mathbf{j}=\mathbf{k},\quad \mathbf{j}\mathbf{k}=\mathbf{i},\quad \mathbf{k}\mathbf{i}=\mathbf{j}.\end{equation*}


Hamilton, when he invented quaternions, he thought of them as of abstract objects obeying simple algebra rules. But now we can realize them as complex matrices. To this end it is enough to define

(7)   \begin{equation*} \mathbf{i}=is_1=\begin{bmatrix}0&i\\i&0\end{bmatrix}, \quad \mathbf{j}=is_2=\begin{bmatrix}0&-1\\1&0\end{bmatrix},\quad \mathbf{k}=is_3=\begin{bmatrix}i&0\\0&-i\end{bmatrix},\end{equation*}

and notice that the algebraic relations defining the multiplication table of quaternions are automatically satisfied! If we would have used Pauli matrices \sigma, we would have to use -i in the above equations – which is not so nice. You would have asked: why minus rather than plus?
Some care is needed, however. One should distinguish the bold \mathbf{i} – the first unit quaternion, from the imaginary complex number i. One should also note that the squares of the three imaginary quaternionic units, in their matrix realization, are -I, that is “minus identity matrix”, instead of just number -1, as in Hamilton’s definition.
In the matrix realization we also have:

(8)   \begin{equation*}\mathbf{i}^*=-\mathbf{i},\quad \mathbf{j}^*=-\mathbf{j},\quad \mathbf{k}^*=-\mathbf{k},\end{equation*}

where the star {}^* denotes Hermitian conjugation.
A general quaternion q is a sum:

    \[q= W+X\mathbf{i}+Y\mathbf{j}+Z\mathbf{k}.\]

In matrix realization it is represented by the matrix

(9)   \begin{equation*}Q=W I+X\mathbf{i}+Y\mathbf{j}+Z\mathbf{k}=\begin{bmatrix}W+iZ&iX-Y\\iX+Y&W-iZ}\end{bmatrix}.\end{equation*}

The conjugated quaternion, q^*= W-X\mathbf{i}-Y\mathbf{j}-Z\mathbf{k} is represented by Hermitian conjugated matrix. Notice that qq^*=q^* q=W^2+X^2+Y^2+Z^2. Unit quaternions are quaternions that have the property that qq^*=1. They are represented by matrices Q such that QQ^*=I, that is by unitary matrices. Moreover we can check that then

    \[\det Q=\det \begin{bmatrix}W+iZ&iX-Y\\iX+Y&W-iZ}\end{bmatrix}=W^2+X^2+Y^2+Z^2=1.\]

Therefore matrices representing unit quaternions are unitary of determinant one. Such matrices form a group, the special unitary group \mathrm{SU}(2). Yet, to dot the i’s and cross the t’s, we still need to show that every 2\times 2 unitary matrix of determinant one is of the form (9). Let therefore U be such a matrix

    \[ U=\begin{bmatrix}a&b\\c&d\end{bmatrix}.\]

Since \det U=ad-bc=1, we can easily check that

    \[\begin{bmatrix}a&b\\c&d\end{bmatrix} \begin{bmatrix}d&-b\\-c&a\end{bmatrix}=I.\]

Therefore

    \[ U^{-1}=\begin{bmatrix}d&-b\\-c&a\end{bmatrix}.\]

On the other hand U is unitary, that means U^*=U^{-1}. But

    \[ U^*=\begin{bmatrix}\bar{a}&\bar{c}\\\bar{b}&\bar{d}\end{bmatrix}.\]

Therefore we must have

    \[\begin{bmatrix}d&-b\\-c&a\end{bmatrix}=\begin{bmatrix}\bar{a}&\bar{c}\\\bar{b}&\bar{d}\end{bmatrix}.\]

That is d=\bar{a},b=-\bar{c}. Writing a=W+iZ, c=Y+iX we get the form (9).

We are interested in rotations in our 3D space. Quantum mechanics suggests using matrices from the group \mathrm{SU}(2). We now know that they are unit quaternions in disguise. But, on the other hand, perhaps quaternions are matrices from \mathrm{SU}(2) in disguise? When Hamilton was inventing quaternions he did not know that he was inventing something that will be useful for rotations. He simply wanted to invent “the algebra of space”. Maxwell first tried to apply quaternions in electrodynamics (for div and curl operators), but that was pretty soon abandoned. Mathematicians at that time were complaining that (see Addendum to the Life of Sir William Rowan Hamilton, LL.D, D.C.L. On Sir W. R. Hamilton’s Irish Descent. On the Calculus of Quaternions. Robert Perceval Graves Dublin: Hodges Figgis, and Co., 1891. ):

As a whole the method is pronounced by most mathematicians to be neither easy nor attractive, the interpretation being hazy or metaphysical and seldom clear and precise.’

Nowadays quaternions are best understood in the framework of Clifford algebras. Unit quaternions are just one example of what are called “spin groups”. We do not need such a general framework here. We will be quite happy with just quaternions and the group \mathrm{SU}(2). But we still have to relate the matrices U from \mathrm{SU}(2) to rotations, and to learn how to use them. Perhaps I will mention it already here, at this point, that the idea is to consider the rotations as points of a certain space. We want to learn about geometry of this space, and we want to investigate different curves in this space. Some of these curves describe rotations and flips of an asymmetric spinning top. Can they be interpreted as “free fall” in this space under some force of gravitation, when gravitation is related to inertia? Perhaps this way we will be able to get a glimpse into otherwise mysterious connection of gravity to quantum mechanics? These questions may sound like somewhat hazy and metaphysical. Therefore let us turn to a good old algebra.

With every vector \vec{v} in our 3D space we can associate Hermitian matrix \vec{v}\cdot\vec{s} of trace zero:

(10)   \begin{equation*}\vec{v}\cdot\vec{s}=v_1s_1+v_2s_2+v_3s_3=\begin{bmatrix}v_3&v_1+iv_2\\v_1-iv_2&-v_3\end{bmatrix}.\end{equation*}

In fact, as it is very easy to see, every Hermitian matrix of trace zero is of this form. Notice that

(11)   \begin{equation*}-\det \vec{v}\cdot\vec{s}=(v_1)^2+(v_2)^2+(v_3)^2=||v||^2.\end{equation*}

Let U be a unitary matrix. We act with unitary matrices on vectors using the “jaw operation” – we take the matrix representation \vec{v}\cdot\vec{s} of the vector in the jaws:

    \[ \vec{v}\cdot\vec{s}\mapsto U\vec{v}\cdot\vec{s}\, U^*.\]

Now, since \vec{v}\cdot\vec{s} is Hermitian and U is unitary, U\vec{v}\cdot\vec{s}\, U^* is also Hermitian. Since \vec{v}\cdot\vec{s} is of trace zero and U is unitary, U\vec{v}\cdot\vec{s}\, U^* is also of trace zero. Therefore there exists vector \tilde{\vec{v}} =(\tilde{v}_1,\tilde{v}_2,\tilde{v}_3) such that:

    \[\tilde{ \vec{v}}\cdot\vec{s}=U\vec{v}\cdot\vec{s}\,U^*.\]

If U\in\mathrm{SU}(2), then \det{U}=\det(U^*)=1. Therefore we must have ||\tilde{\vec{v}}||=||\vec{v}||. In other words the transformation (which is linear) \vec{v}\mapsto \tilde{\vec{v}} is an isometry – it preserves the length of all vectors. And it is known that every isometry (that maps 0 into 0, which is the case) is a rotation (see Wikipedia, Isometry). Therefore there exists a unique rotation matrix R such that \tilde{\vec{v}}=R\vec{v}. This way we have a map from \mathrm{SU}(2) to O(3) – the group of orthogonal matrices. We do not know yet if \det(R)=1, which is the case, but it needs a proof, some proof….