Lie Groups II March 9, 2025
Adjoint Representation and Campbell Identity ¶ We define A d X ( Y ) \mathrm{Ad}_{X}(Y) Ad X ( Y ) by:
A d X ( Y ) = X Y X − 1 \mathrm{Ad}_{X}(Y) = X Y X^{-1} Ad X ( Y ) = X Y X − 1 and define a matrix function f ( t ) f(t) f ( t ) by:
f ( t ) = A d e t X ( Y ) = e t X Y e − t X . f(t) = \mathrm{Ad}_{e^{tX}}(Y) = e^{tX} Y e^{-tX}\,. f ( t ) = Ad e tX ( Y ) = e tX Y e − tX . Differentiate f ( t ) f(t) f ( t ) :
f ′ ( t ) = d d t A d e t X ( Y ) = d d t e t X Y e − t X = f'(t) = {d\over d t} \mathrm{Ad}_{e^{tX}}(Y)
= {d\over d t} e^{tX} Y e^{-tX} = f ′ ( t ) = d t d Ad e tX ( Y ) = d t d e tX Y e − tX = = X e t X Y e − t X − e t X Y X e − t X = = X e^{tX} Y e^{-tX} - e^{tX} Y X e^{-tX} = = X e tX Y e − tX − e tX Y X e − tX = = e t X X Y e − t X − e t X Y X e − t X = = e^{tX} X Y e^{-tX} - e^{tX} Y X e^{-tX} = = e tX X Y e − tX − e tX Y X e − tX = = e t X ( X Y − Y X ) e − t X = = e^{tX} (XY - YX) e^{-tX} = = e tX ( X Y − Y X ) e − tX = = e t X [ X , Y ] e − t X . = e^{tX} [X, Y] e^{-tX}\,. = e tX [ X , Y ] e − tX . We apply it recursively for the second derivative:
f ′ ′ ( t ) = d 2 d t 2 A d e t X ( Y ) = d 2 d t 2 e t X Y e − t X = f''(t) = {d^2\over d t^2} \mathrm{Ad}_{e^{tX}}(Y)
= {d^2\over d t^2} e^{tX} Y e^{-tX} = f ′′ ( t ) = d t 2 d 2 Ad e tX ( Y ) = d t 2 d 2 e tX Y e − tX = = d d t e t X [ X , Y ] e − t X = = {d\over d t} e^{tX} [X, Y] e^{-tX} = = d t d e tX [ X , Y ] e − tX = = e t X [ X , [ X , Y ] ] e − t X . = e^{tX} [X, [X, Y]] e^{-tX}\,. = e tX [ X , [ X , Y ]] e − tX . We evaluate these derivatives at t = 0 t=0 t = 0 :
f ′ ( 0 ) = [ X , Y ] , f'(0)=[X, Y]\,, f ′ ( 0 ) = [ X , Y ] ,
f ′ ′ ( 0 ) = [ X , [ X , Y ] ] . f''(0)=[X, [X, Y]]\,. f ′′ ( 0 ) = [ X , [ X , Y ]] . We expand f ( t ) f(t) f ( t ) using a Taylor series around the point t = 0 t=0 t = 0 :
f ( t ) = f ( 0 ) + 1 1 ! f ′ ( 0 ) t + 1 2 ! f ′ ′ ( 0 ) t 2 + 1 3 ! f ′ ′ ′ ( 0 ) t 3 + ⋯ f(t)= f(0)
+ {1\over 1!}f'(0) t
+ {1\over 2!}f''(0) t^2
+ {1\over 3!}f'''(0) t^3
+ \cdots f ( t ) = f ( 0 ) + 1 ! 1 f ′ ( 0 ) t + 2 ! 1 f ′′ ( 0 ) t 2 + 3 ! 1 f ′′′ ( 0 ) t 3 + ⋯ and we get:
f ( t ) = e t X Y e − t X = Y + 1 1 ! [ X , Y ] t + 1 2 ! [ X , [ X , Y ] ] t 2 + 1 3 ! [ X , [ X , [ X , Y ] ] ] t 3 + ⋯ f(t) = e^{tX} Y e^{-tX} = Y
+ {1\over 1!} [X,Y] t
+ {1\over 2!} [X,[X,Y]] t^2
+ {1\over 3!} [X,[X,[X,Y]]] t^3
+ \cdots f ( t ) = e tX Y e − tX = Y + 1 ! 1 [ X , Y ] t + 2 ! 1 [ X , [ X , Y ]] t 2 + 3 ! 1 [ X , [ X , [ X , Y ]]] t 3 + ⋯ We can set t = 1 t=1 t = 1 :
f ( 1 ) = e X Y e − X = Y + 1 1 ! [ X , Y ] + 1 2 ! [ X , [ X , Y ] ] + 1 3 ! [ X , [ X , [ X , Y ] ] ] + ⋯ f(1) = e^{X} Y e^{-X} = Y
+ {1\over 1!} [X,Y]
+ {1\over 2!} [X,[X,Y]]
+ {1\over 3!} [X,[X,[X,Y]]]
+ \cdots f ( 1 ) = e X Y e − X = Y + 1 ! 1 [ X , Y ] + 2 ! 1 [ X , [ X , Y ]] + 3 ! 1 [ X , [ X , [ X , Y ]]] + ⋯ This last equation is called the Campbell identity. We can also write it as:
e − X Y e X = Y + 1 1 ! [ − X , Y ] + 1 2 ! [ − X , [ − X , Y ] ] + 1 3 ! [ − X , [ − X , [ − X , Y ] ] ] + ⋯ = e^{-X} Y e^{X} = Y
+ {1\over 1!} [-X,Y]
+ {1\over 2!} [-X,[-X,Y]]
+ {1\over 3!} [-X,[-X,[-X,Y]]]
+ \cdots = e − X Y e X = Y + 1 ! 1 [ − X , Y ] + 2 ! 1 [ − X , [ − X , Y ]] + 3 ! 1 [ − X , [ − X , [ − X , Y ]]] + ⋯ = = Y + 1 1 ! [ Y , X ] + 1 2 ! [ [ Y , X ] , X ] + 1 3 ! [ [ [ Y , X ] , X ] , X ] + ⋯ = Y
+ {1\over 1!} [Y,X]
+ {1\over 2!} [[Y,X],X]
+ {1\over 3!} [[[Y,X],X],X]
+ \cdots = Y + 1 ! 1 [ Y , X ] + 2 ! 1 [[ Y , X ] , X ] + 3 ! 1 [[[ Y , X ] , X ] , X ] + ⋯ Lie Algebra ¶ Lie algebra g g g for a Lie group G G G that is a topologically closed subgroup of
GL(n, R) is defined by:
g = { X : exp ( t X ) ∈ G ∀ t ∈ R } . g = \{X: \exp(t X) \in G \quad \forall t \in \mathbb{R}\}\,. g = { X : exp ( tX ) ∈ G ∀ t ∈ R } . The following can be proven:
g g g is a vector space
X , Y ∈ g ⇒ [ X , Y ] ∈ g X, Y \in g \Rightarrow [X, Y] \in g X , Y ∈ g ⇒ [ X , Y ] ∈ g
g g g is the tangent space to G G G at I I I .
exp : g → G \exp: g \rightarrow G exp : g → G is locally invertible
Invariant subgroup a a a of a group g g g : for every a a a and
g g g the following holds:
g a g − 1 = a ′ , g a g^{-1} = a'\,, g a g − 1 = a ′ , where a ′ a' a ′ is a member of a a a .
Invariant abelian subgroup : is an invariant subgroup that is abelian: a a ′ = a ′ a a a' =
a' a a a ′ = a ′ a , or [ a , a ′ ] = 0 [a, a'] = 0 [ a , a ′ ] = 0 .
Lie group is:
Typical examples of semisimple groups are the direct products of simple groups.
Using the structure constants [ L i , L j ] = C i j k L k [L_i, L_j] = C_{ij}{}^k L_k [ L i , L j ] = C ij k L k we can define a
metric tensor (also called Killing form ):
g i j = g j i = C i k l C j l k . g_{ij} = g_{ji} = C_{ik}{}^l C_{jl}{}^k\,. g ij = g ji = C ik l C j l k . The regular representation is:
( L i ) j k = C i j k . (L_i)_{j}{}^k = C_{ij}{}^k\,. ( L i ) j k = C ij k . From this it follows:
g i j = C i k l C j l k = ( L i ) k l ( L j ) l k = T r ( L i , L j ) ≡ ( L i , L j ) . g_{ij} = C_{ik}{}^l C_{jl}{}^k = (L_i)_{k}{}^l (L_j)_{l}{}^k
= \mathrm{Tr} (L_i, L_j) \equiv (L_i, L_j)\,. g ij = C ik l C j l k = ( L i ) k l ( L j ) l k = Tr ( L i , L j ) ≡ ( L i , L j ) . This fulfills all properties of a metric, e.g.:
( L i , L j ) = ( L j , L i ) , (L_i,L_j) = (L_j, L_i)\,, ( L i , L j ) = ( L j , L i ) ,
( L i + L j , L k ) = ( L i , L k ) + ( L j , L k ) , (L_i+L_j,L_k) = (L_i, L_k) + (L_j, L_k)\,, ( L i + L j , L k ) = ( L i , L k ) + ( L j , L k ) , but it is not necessarily positive definite (i.e., ( L i , L i ) > 0 (L_i, L_i) > 0 ( L i , L i ) > 0 ).
Theorem: a Lie algebra is semisimple if and only if det g i j ≠ 0 \det g_{ij} \neq 0 det g ij = 0 .
The center of a group G is a set of elements that commute with every element of
G:
Z ( G ) = { z ∈ G ∣ ∀ g ∈ G , z g = g z } . Z(G) = \{z ∈ G \quad | \quad ∀g ∈ G, \quad zg = gz\}\,. Z ( G ) = { z ∈ G ∣ ∀ g ∈ G , z g = g z } . The center is a normal subgroup of G, so the quotient G / Z ( G ) G / Z(G) G / Z ( G ) is a group.
To construct every other connected Lie group with the same Lie algebra su(2) as
SU(2), such as SO(3), we take quotients of SU(2) by discrete subgroups of its
center Z ( S U ( 2 ) ) = { I , − I } Z(SU(2))=\{I,−I\} Z ( S U ( 2 )) = { I , − I } . The possible quotients are:
S U ( 2 ) / { I } ≅ S U ( 2 ) , SU(2)/\{I\}≅SU(2), S U ( 2 ) / { I } ≅ S U ( 2 ) , S U ( 2 ) / { I , − I } ≅ S O ( 3 ) . SU(2)/\{I,−I\}≅SO(3). S U ( 2 ) / { I , − I } ≅ SO ( 3 ) . Utilities ¶ Computing exp ( θ J 3 ) P 1 exp ( − θ J 3 ) \exp(\theta J_3) P_1 \exp(-\theta J_3) exp ( θ J 3 ) P 1 exp ( − θ J 3 ) ¶ Where J i J_i J i are generators of the SO(3) group and P 1 P_1 P 1 are generators of the
translation group. We do not know their matrices, we only know their
commutators. We use the formula above with
t = θ t = \theta t = θ , X = J 3 X=J_3 X = J 3 and Y = P 1 Y=P_1 Y = P 1 :
e θ J 3 P 1 e − θ J 3 = P 1 + 1 1 ! [ J 3 , P 1 ] θ + 1 2 ! [ J 3 , [ J 3 , P 1 ] ] θ 2 + 1 3 ! [ J 3 , [ J 3 , [ J 3 , P 1 ] ] ] θ 3 + ⋯ . e^{\theta J_3} P_1 e^{-\theta J_3} = P_1
+ {1\over 1!} [J_3,P_1] \theta
+ {1\over 2!} [J_3,[J_3,P_1]] \theta^2
+ {1\over 3!} [J_3,[J_3,[J_3,P_1]]] \theta^3
+ \cdots\,. e θ J 3 P 1 e − θ J 3 = P 1 + 1 ! 1 [ J 3 , P 1 ] θ + 2 ! 1 [ J 3 , [ J 3 , P 1 ]] θ 2 + 3 ! 1 [ J 3 , [ J 3 , [ J 3 , P 1 ]]] θ 3 + ⋯ . We need to compute the commutators. The Lie algebra commutator is:
[ J i , P j ] = ϵ i j k P k . [J_i, P_j] = \epsilon_{ijk} P_k\,. [ J i , P j ] = ϵ ijk P k . So:
[ J 3 , P 1 ] = ϵ 312 P 2 = P 2 , [J_3, P_1] = \epsilon_{312} P_2 = P_2\,, [ J 3 , P 1 ] = ϵ 312 P 2 = P 2 ,
[ J 3 , [ J 3 , P 1 ] ] = [ J 3 , P 2 ] = ϵ 321 P 1 = − P 1 , [J_3, [J_3, P_1]] = [J_3, P_2] = \epsilon_{321} P_1 = -P_1\,, [ J 3 , [ J 3 , P 1 ]] = [ J 3 , P 2 ] = ϵ 321 P 1 = − P 1 ,
[ J 3 , [ J 3 , [ J 3 , P 1 ] ] ] = [ J 3 , − P 1 ] = − ϵ 312 P 2 = − P 2 . [J_3, [J_3, [J_3, P_1]]] = [J_3, -P_1] = -\epsilon_{312} P_2 = -P_2\,. [ J 3 , [ J 3 , [ J 3 , P 1 ]]] = [ J 3 , − P 1 ] = − ϵ 312 P 2 = − P 2 . And higher terms repeat. The commutators alternate between P 2 P_2 P 2 and P 1 P_1 P 1 and
their signs:
e θ J 3 P 1 e − θ J 3 = P 1 + 1 1 ! P 2 θ + 1 2 ! ( − P 1 ) θ 2 + 1 3 ! ( − P 2 ) θ 3 + 1 4 ! P 1 θ 4 + 1 5 ! P 2 θ 5 + ⋯ = e^{\theta J_3} P_1 e^{-\theta J_3} = P_1
+ {1\over 1!} P_2 \theta
+ {1\over 2!} (-P_1) \theta^2
+ {1\over 3!} (-P_2) \theta^3
+ {1\over 4!} P_1 \theta^4
+ {1\over 5!} P_2 \theta^5
+ \cdots = e θ J 3 P 1 e − θ J 3 = P 1 + 1 ! 1 P 2 θ + 2 ! 1 ( − P 1 ) θ 2 + 3 ! 1 ( − P 2 ) θ 3 + 4 ! 1 P 1 θ 4 + 5 ! 1 P 2 θ 5 + ⋯ = = P 1 ( 1 − 1 2 ! θ 2 + 1 4 ! θ 4 + ⋯ ) + P 2 ( 1 1 ! θ − 1 3 ! θ 3 + 1 5 ! θ 5 + ⋯ ) = = P_1 (1 - {1\over 2!} \theta^2 + {1\over 4!} \theta^4 + \cdots)
+ P_2 ({1\over 1!} \theta - {1\over 3!} \theta^3 + {1\over 5!} \theta^5 + \cdots) = = P 1 ( 1 − 2 ! 1 θ 2 + 4 ! 1 θ 4 + ⋯ ) + P 2 ( 1 ! 1 θ − 3 ! 1 θ 3 + 5 ! 1 θ 5 + ⋯ ) = = P 1 cos ( θ ) + P 2 sin ( θ ) . = P_1 \cos(\theta) + P_2 \sin(\theta)\,. = P 1 cos ( θ ) + P 2 sin ( θ ) . Any vector operator has the same Lie bracket, for example for J i J_i J i we have
[ J i , J j ] = ϵ i j k J k [J_i, J_j] = \epsilon_{ijk} J_k [ J i , J j ] = ϵ ijk J k and obtain:
e θ J 3 J 1 e − θ J 3 = J 1 cos ( θ ) + J 2 sin ( θ ) . e^{\theta J_3} J_1 e^{-\theta J_3} = J_1 \cos(\theta) + J_2 \sin(\theta)\,. e θ J 3 J 1 e − θ J 3 = J 1 cos ( θ ) + J 2 sin ( θ ) . Computing exp ( θ J 3 ) P 2 exp ( − θ J 3 ) \exp(\theta J_3) P_2 \exp(-\theta J_3) exp ( θ J 3 ) P 2 exp ( − θ J 3 ) ¶ Similarly, let’s compute:
e θ J 3 P 2 e − θ J 3 = P 2 + 1 1 ! [ J 3 , P 2 ] θ + 1 2 ! [ J 3 , [ J 3 , P 2 ] ] θ 2 + 1 3 ! [ J 3 , [ J 3 , [ J 3 , P 2 ] ] ] θ 3 + ⋯ = e^{\theta J_3} P_2 e^{-\theta J_3} = P_2
+ {1\over 1!} [J_3,P_2] \theta
+ {1\over 2!} [J_3,[J_3,P_2]] \theta^2
+ {1\over 3!} [J_3,[J_3,[J_3,P_2]]] \theta^3
+ \cdots = e θ J 3 P 2 e − θ J 3 = P 2 + 1 ! 1 [ J 3 , P 2 ] θ + 2 ! 1 [ J 3 , [ J 3 , P 2 ]] θ 2 + 3 ! 1 [ J 3 , [ J 3 , [ J 3 , P 2 ]]] θ 3 + ⋯ = = P 2 + 1 1 ! ( − P 1 ) θ + 1 2 ! ( − P 2 ) θ 2 + 1 3 ! P 1 θ 3 + 1 4 ! P 2 θ 4 + 1 5 ! ( − P 1 ) θ 5 + ⋯ = = P_2
+ {1\over 1!} (-P_1) \theta
+ {1\over 2!} (-P_2) \theta^2
+ {1\over 3!} P_1 \theta^3
+ {1\over 4!} P_2 \theta^4
+ {1\over 5!} (-P_1) \theta^5
+ \cdots = = P 2 + 1 ! 1 ( − P 1 ) θ + 2 ! 1 ( − P 2 ) θ 2 + 3 ! 1 P 1 θ 3 + 4 ! 1 P 2 θ 4 + 5 ! 1 ( − P 1 ) θ 5 + ⋯ = = P 1 ( − θ 1 ! + θ 3 3 ! − θ 5 5 ! + ⋯ ) + P 2 ( 1 − θ 2 2 ! + θ 4 4 ! + ⋯ ) = = P_1 \left(-{\theta\over 1!} + {\theta^3\over 3!} - {\theta^5\over 5!} + \cdots\right)
+ P_2 \left(1 - {\theta^2\over 2!} + {\theta^4\over 4!} + \cdots \right) = = P 1 ( − 1 ! θ + 3 ! θ 3 − 5 ! θ 5 + ⋯ ) + P 2 ( 1 − 2 ! θ 2 + 4 ! θ 4 + ⋯ ) = = − P 1 sin ( θ ) + P 2 cos ( θ ) . = -P_1 \sin(\theta) + P_2 \cos(\theta)\,. = − P 1 sin ( θ ) + P 2 cos ( θ ) . Computing exp ( θ J 3 ) P 3 exp ( − θ J 3 ) \exp(\theta J_3) P_3 \exp(-\theta J_3) exp ( θ J 3 ) P 3 exp ( − θ J 3 ) ¶ Finally:
e θ J 3 P 3 e − θ J 3 = P 3 + 1 1 ! [ J 3 , P 3 ] θ + 1 2 ! [ J 3 , [ J 3 , P 3 ] ] θ 2 + ⋯ = P 3 , e^{\theta J_3} P_3 e^{-\theta J_3} = P_3
+ {1\over 1!} [J_3,P_3] \theta
+ {1\over 2!} [J_3,[J_3,P_3]] \theta^2
+ \cdots = P_3\,, e θ J 3 P 3 e − θ J 3 = P 3 + 1 ! 1 [ J 3 , P 3 ] θ + 2 ! 1 [ J 3 , [ J 3 , P 3 ]] θ 2 + ⋯ = P 3 , since all the commutators are zero.
Summary ¶ We have thus shown that for any operator satisfying
[ J i , P j ] = ϵ i j k P k [J_i, P_j] = \epsilon_{ijk} P_k [ J i , P j ] = ϵ ijk P k we get:
e θ J 3 P 1 e − θ J 3 = P 1 cos ( θ ) + P 2 sin ( θ ) , e θ J 3 P 2 e − θ J 3 = − P 1 sin ( θ ) + P 2 cos ( θ ) , e θ J 3 P 3 e − θ J 3 = P 3 . \begin{aligned}
e^{\theta J_3} P_1 e^{-\theta J_3} &= P_1 \cos(\theta) + P_2 \sin(\theta)\,, \\
e^{\theta J_3} P_2 e^{-\theta J_3} &= -P_1 \sin(\theta) + P_2 \cos(\theta)\,, \\
e^{\theta J_3} P_3 e^{-\theta J_3} &= P_3\,.
\end{aligned} e θ J 3 P 1 e − θ J 3 e θ J 3 P 2 e − θ J 3 e θ J 3 P 3 e − θ J 3 = P 1 cos ( θ ) + P 2 sin ( θ ) , = − P 1 sin ( θ ) + P 2 cos ( θ ) , = P 3 . We can thus write these as a single equation in index notation:
e θ J 3 P j e − θ J 3 = ( R 3 ) i j P i , e^{\theta J_3} P_j e^{-\theta J_3}
=(R_3)^i{}_j P_i\,, e θ J 3 P j e − θ J 3 = ( R 3 ) i j P i , where the matrix ( R 3 ) i j (R_3)^i{}_j ( R 3 ) i j is a rotation around the z z z -axis:
( R 3 ( θ ) ) i j = ( cos θ − sin θ 0 sin θ cos θ 0 0 0 1 ) . (R_3(\theta))^i{}_j
=\begin{pmatrix}
\cos \theta &-\sin \theta & 0 \\
\sin \theta & \cos \theta & 0 \\
0 & 0 & 1 \\
\end{pmatrix}\,. ( R 3 ( θ ) ) i j = ⎝ ⎛ cos θ sin θ 0 − sin θ cos θ 0 0 0 1 ⎠ ⎞ . We are interested in computing a ⋅ P = a j P j {\bf a} \cdot {\bf P} = a^j P_j a ⋅ P = a j P j :
e θ J 3 a j P j e − θ J 3 = ( R 3 ) i j a j P i e^{\theta J_3} a^j P_j e^{-\theta J_3}
=(R_3)^i{}_j a^j P_i e θ J 3 a j P j e − θ J 3 = ( R 3 ) i j a j P i and using the below indentity:
e θ J 3 e a j P j e − θ J 3 = e e θ J 3 a j P j e − θ J 3 = e ( R 3 ) i j a j P i , e^{\theta J_3} e^{a^j P_j} e^{-\theta J_3}
=e^{e^{\theta J_3} a^j P_j e^{-\theta J_3}}
=e^{(R_3)^i{}_j a^j P_i}\,, e θ J 3 e a j P j e − θ J 3 = e e θ J 3 a j P j e − θ J 3 = e ( R 3 ) i j a j P i , or:
e θ J 3 T ( a ) e − θ J 3 = T ( R 3 ( θ ) a ) . e^{\theta J_3} T({\bf a}) e^{-\theta J_3} =T(R_3(\theta) {\bf a})\,. e θ J 3 T ( a ) e − θ J 3 = T ( R 3 ( θ ) a ) . In an analogous way we can obtain:
e θ J 1 T ( a ) e − θ J 1 = T ( R 1 ( θ ) a ) , e^{\theta J_1} T({\bf a}) e^{-\theta J_1} =T(R_1(\theta) {\bf a})\,, e θ J 1 T ( a ) e − θ J 1 = T ( R 1 ( θ ) a ) , e θ J 2 T ( a ) e − θ J 2 = T ( R 2 ( θ ) a ) . e^{\theta J_2} T({\bf a}) e^{-\theta J_2} =T(R_2(\theta) {\bf a})\,. e θ J 2 T ( a ) e − θ J 2 = T ( R 2 ( θ ) a ) . Putting it all together we get:
e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 T ( a ) e − θ 3 J 3 e − θ 2 J 2 e − θ 1 J 1 = e^{\theta_1 J_1} e^{\theta_2 J_2} e^{\theta_3 J_3}
T({\bf a})
e^{-\theta_3 J_3} e^{-\theta_2 J_2} e^{-\theta_1 J_1}= e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 T ( a ) e − θ 3 J 3 e − θ 2 J 2 e − θ 1 J 1 = = e θ 1 J 1 e θ 2 J 2 T ( R 3 ( θ 3 ) a ) e − θ 2 J 2 e − θ 1 J 1 = = e^{\theta_1 J_1} e^{\theta_2 J_2}
T(R_3(\theta_3) {\bf a})
e^{-\theta_2 J_2} e^{-\theta_1 J_1}= = e θ 1 J 1 e θ 2 J 2 T ( R 3 ( θ 3 ) a ) e − θ 2 J 2 e − θ 1 J 1 = = e θ 1 J 1 T ( R 2 ( θ 2 ) R 3 ( θ 3 ) a ) e − θ 1 J 1 = e^{\theta_1 J_1} T(R_2(\theta_2) R_3(\theta_3) {\bf a}) e^{-\theta_1 J_1} = e θ 1 J 1 T ( R 2 ( θ 2 ) R 3 ( θ 3 ) a ) e − θ 1 J 1 = T ( R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) a ) . = T(R_1(\theta_1) R_2(\theta_2) R_3(\theta_3) {\bf a}) \,. = T ( R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) a ) . In other words, we can see how a general element of the rotation group
e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 e^{\theta_1 J_1} e^{\theta_2 J_2} e^{\theta_3 J_3} e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 acts on a general element
of the translation group T ( a ) T({\bf a}) T ( a ) . The result is another translation
T ( R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) a ) T(R_1(\theta_1) R_2(\theta_2) R_3(\theta_3) {\bf a}) T ( R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) a ) and the translation
vector a {\bf a} a gets rotated with the three rotation matrices
a → R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) a {\bf a} \rightarrow R_1(\theta_1) R_2(\theta_2) R_3(\theta_3) {\bf a} a → R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) a .
For T=J ¶ As a special case, we’ll use P i = J i P_i = J_i P i = J i , and we’ll use n \bf n n instead of a \bf
a a :
e θ J 3 n j J j e − θ J 3 = ( R 3 ) i j n j J i e^{\theta J_3} n^j J_j e^{-\theta J_3}
=(R_3)^i{}_j n^j J_i e θ J 3 n j J j e − θ J 3 = ( R 3 ) i j n j J i In an analogous way we can obtain:
e θ J 1 n j J j e − θ J 1 = ( R 1 ) i j n j J i e^{\theta J_1} n^j J_j e^{-\theta J_1}
=(R_1)^i{}_j n^j J_i e θ J 1 n j J j e − θ J 1 = ( R 1 ) i j n j J i e θ J 2 n j J j e − θ J 2 = ( R 2 ) i j n j J i e^{\theta J_2} n^j J_j e^{-\theta J_2}
=(R_2)^i{}_j n^j J_i e θ J 2 n j J j e − θ J 2 = ( R 2 ) i j n j J i Putting it all together we get:
e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 n j J j e − θ 3 J 3 e − θ 2 J 2 e − θ 1 J 1 = e^{\theta_1 J_1} e^{\theta_2 J_2} e^{\theta_3 J_3}
n^j J_j
e^{-\theta_3 J_3} e^{-\theta_2 J_2} e^{-\theta_1 J_1}= e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 n j J j e − θ 3 J 3 e − θ 2 J 2 e − θ 1 J 1 = = e θ 1 J 1 e θ 2 J 2 ( R 3 ) i j n j J i e − θ 2 J 2 e − θ 1 J 1 = = e^{\theta_1 J_1} e^{\theta_2 J_2}
(R_3)^i{}_j n^j J_i
e^{-\theta_2 J_2} e^{-\theta_1 J_1}= = e θ 1 J 1 e θ 2 J 2 ( R 3 ) i j n j J i e − θ 2 J 2 e − θ 1 J 1 = = e θ 1 J 1 ( R 2 ) i j ( R 3 ) j k n k J i e − θ 1 J 1 = = e^{\theta_1 J_1}
(R_2)^i{}_j (R_3)^j{}_k n^k J_i
e^{-\theta_1 J_1} = = e θ 1 J 1 ( R 2 ) i j ( R 3 ) j k n k J i e − θ 1 J 1 = = ( R 1 ) i j ( R 2 ) j k ( R 3 ) k l n l J i . = (R_1)^i{}_j (R_2)^j{}_k (R_3)^k{}_l n^l J_i \,. = ( R 1 ) i j ( R 2 ) j k ( R 3 ) k l n l J i . This shows how to “encode” the 3D vector n \bf n n into any j j j representation.
The RHS is encoding the rotated vector, the LHS is encoding the original vector
and then transforming it using the “sandwich” formula. Thus showing that the
“sandwich” transformation is equivalent to a rotation. It is thus showing how
any SU(2) transformation corresponds to SO(3) transformation, the “minus”
cancels out in the sandwich, so 360 degree rotation transforms n \bf n n into
itself. To obtain the 360 rotation equal to -1 we need to act with just the
left part of the sandwich on some “spinor”, and it would transform it to -1.
Thus we have proven:
U ( θ 1 , θ 2 , θ 3 ) a ⋅ J U ( θ 1 , θ 2 , θ 3 ) − 1 = ( R ( θ 1 , θ 2 , θ 3 ) a ) ⋅ J , U(\theta_1, \theta_2, \theta_3)
\ {\bf a} \cdot {\bf J}\
U(\theta_1, \theta_2, \theta_3)^{-1}
=
(R(\theta_1, \theta_2, \theta_3) {\bf a}) \cdot {\bf J}\,, U ( θ 1 , θ 2 , θ 3 ) a ⋅ J U ( θ 1 , θ 2 , θ 3 ) − 1 = ( R ( θ 1 , θ 2 , θ 3 ) a ) ⋅ J , where:
U ( θ 1 , θ 2 , θ 3 ) = e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 , U(\theta_1, \theta_2, \theta_3)
=e^{\theta_1 J_1} e^{\theta_2 J_2} e^{\theta_3 J_3}\,, U ( θ 1 , θ 2 , θ 3 ) = e θ 1 J 1 e θ 2 J 2 e θ 3 J 3 , R ( θ 1 , θ 2 , θ 3 ) = R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) . R(\theta_1, \theta_2, \theta_3)
= R_1(\theta_1) R_2(\theta_2) R_3(\theta_3) \,. R ( θ 1 , θ 2 , θ 3 ) = R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) . Spin 1/2 ¶ For spin j=1/2 we have:
J i = − i 1 2 σ i J_i = -i {1\over 2} \sigma_i J i = − i 2 1 σ i Now we would like to show that:
U ( θ 1 , θ 2 , θ 3 ) = e − i θ 1 2 σ 1 e − i θ 2 2 σ 2 e − i θ 3 2 σ 3 = U ( n , θ ) = e − i θ 2 ( n ⋅ σ ) , U(\theta_1, \theta_2, \theta_3)
=e^{-i{\theta_1\over2} \sigma_1} e^{-i{\theta_2\over2} \sigma_2} e^{-i{\theta_3\over2} \sigma_3}
=U({\bf n}, \theta) = e^{-i {\theta\over2}\left({\bf n}\cdot {\bf\sigma} \right)}\,, U ( θ 1 , θ 2 , θ 3 ) = e − i 2 θ 1 σ 1 e − i 2 θ 2 σ 2 e − i 2 θ 3 σ 3 = U ( n , θ ) = e − i 2 θ ( n ⋅ σ ) , where n {\bf n} n is a unit vector representing the axis of rotation and θ \theta θ
is the rotation around this axis. This follows from the fact that the LHS is a
general element of the SU(2) group, and so is the RHS. Constructively, this is
proven in a separate file.
TODO: We also want to prove that the three rotation matrices can be
written as just one matrix:
R ( θ 1 , θ 2 , θ 3 ) = R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) = R ( n , θ ) . R(\theta_1, \theta_2, \theta_3)
= R_1(\theta_1) R_2(\theta_2) R_3(\theta_3)
= R({\bf n}, \theta)\,. R ( θ 1 , θ 2 , θ 3 ) = R 1 ( θ 1 ) R 2 ( θ 2 ) R 3 ( θ 3 ) = R ( n , θ ) . Then we get:
U ( n , θ ) a ⋅ σ U ( n , θ ) − 1 = ( R ( n , θ ) a ) ⋅ σ , U({\bf n}, \theta)
\ {\bf a} \cdot {\bf \sigma}\
U({\bf n}, \theta)^{-1}
=
(R({\bf n}, \theta) {\bf a}) \cdot {\bf \sigma}\,, U ( n , θ ) a ⋅ σ U ( n , θ ) − 1 = ( R ( n , θ ) a ) ⋅ σ , Here U U U can be expanded as:
U ( n , θ ) = e − i θ 2 ( n ⋅ σ ) = cos ( θ 2 ) − i sin ( θ 2 ) ( n ⋅ σ ) . U({\bf n}, \theta) = e^{-i {\theta\over2}\left({\bf n}\cdot {\bf\sigma} \right)}
= \cos\left({\theta\over 2}\right) - i \sin\left({\theta\over2}\right) \left({\bf n}\cdot {\bf\sigma} \right)\,. U ( n , θ ) = e − i 2 θ ( n ⋅ σ ) = cos ( 2 θ ) − i sin ( 2 θ ) ( n ⋅ σ ) . We can now identify this relation with quaternions as follows:
U ( n , θ ) = cos ( θ 2 ) + sin ( θ 2 ) ( n 1 ( − i σ 1 ) + n 2 ( − i σ 2 ) + n 3 ( − i σ 3 ) ) , U({\bf n}, \theta)
= \cos\left({\theta\over 2}\right) + \sin\left({\theta\over2}\right)
\left(n^1 (-i\sigma_1) + n^2 (-i\sigma_2) + n^3 (-i\sigma_3) \right)\,, U ( n , θ ) = cos ( 2 θ ) + sin ( 2 θ ) ( n 1 ( − i σ 1 ) + n 2 ( − i σ 2 ) + n 3 ( − i σ 3 ) ) , q = cos ( θ 2 ) + sin ( θ 2 ) ( n 1 i + n 2 j + n 3 k ) . q
= \cos\left({\theta\over 2}\right) + \sin\left({\theta\over2}\right)
\left(n^1 i + n^2 j + n^3 k \right)\,. q = cos ( 2 θ ) + sin ( 2 θ ) ( n 1 i + n 2 j + n 3 k ) . And the rotation equation is:
q ( a 1 i + a 2 j + a 3 k ) q − 1 = ( R ( q ) a ) ⋅ ( i , j , k ) q (a^1 i + a^2 j + a^3 k) q^{-1} = (R(q){\bf a}) \cdot (i, j, k) q ( a 1 i + a 2 j + a 3 k ) q − 1 = ( R ( q ) a ) ⋅ ( i , j , k ) In other words, the LHS is encoding the vector a {\bf a} a into a quaternion as
a 1 i + a 2 j + a 3 k a^1 i + a^2 j + a^3 k a 1 i + a 2 j + a 3 k , and the RHS gives us the rotated quaternion vector
that we can decode back.
Identity g e X g − 1 = e g X g − 1 g e^X g^{-1} = e^{g X g^{-1}} g e X g − 1 = e g X g − 1 ¶ The identity
g e X g − 1 = e g X g − 1 g e^X g^{-1} = e^{g X g^{-1}} g e X g − 1 = e g X g − 1 can be proven using:
g e X g − 1 = g ( ∑ n = 0 ∞ X n n ! ) g − 1 = ∑ n = 0 ∞ g X n g − 1 n ! = ∑ n = 0 ∞ ( g X g − 1 ) n n ! = e g X g − 1 . g e^X g^{-1}=g \left( \sum_{n=0}^\infty {X^n\over n!} \right) g^{-1}
=\sum_{n=0}^\infty {g X^n g^{-1} \over n!}
=\sum_{n=0}^\infty {(g X g^{-1})^n \over n!}
=e^{g X g^{-1}}\,. g e X g − 1 = g ( n = 0 ∑ ∞ n ! X n ) g − 1 = n = 0 ∑ ∞ n ! g X n g − 1 = n = 0 ∑ ∞ n ! ( g X g − 1 ) n = e g X g − 1 . Computing exp ( θ 3 J 3 ) exp ( θ 1 P 1 ) exp ( − θ 3 J 3 ) \exp(\theta_3 J_3) \exp(\theta_1 P_1) \exp(-\theta_3 J_3) exp ( θ 3 J 3 ) exp ( θ 1 P 1 ) exp ( − θ 3 J 3 ) ¶ Using the identity g e X g − 1 = e g X g − 1 g e^X g^{-1} = e^{g X g^{-1}} g e X g − 1 = e g X g − 1 we get:
e θ 3 J 3 e θ 1 P 1 e − θ 3 J 3 = exp ( e θ 3 J 3 ( θ 1 P 1 ) e − θ 3 J 3 ) = e θ 1 ( P 1 cos θ 3 + P 2 sin θ 3 ) . e^{\theta_3 J_3} e^{\theta_1 P_1} e^{-\theta_3 J_3}
=\exp\left(e^{\theta_3 J_3} (\theta_1 P_1) e^{-\theta_3 J_3}\right)
=e^{\theta_1 \left(P_1 \cos\theta_3 + P_2 \sin\theta_3\right)} \,. e θ 3 J 3 e θ 1 P 1 e − θ 3 J 3 = exp ( e θ 3 J 3 ( θ 1 P 1 ) e − θ 3 J 3 ) = e θ 1 ( P 1 c o s θ 3 + P 2 s i n θ 3 ) . If v \mathbf{v} v is a vector in R^3 and k \mathbf{k} k is a unit vector (axis of
rotation) and θ \theta θ the angle of rotation, then the vector v \mathbf{v} v gets
rotated to v r o t \mathbf{v}_\mathrm{rot} v rot .
We define a matrix K \mathbf{K} K using:
K i j = − ϵ i j k k k = ( 0 − k 3 k 2 k 3 0 − k 1 − k 2 k 1 0 ) \mathbf{K}_{ij} = -\epsilon_{ijk} k^k
= \begin{pmatrix}
0 & -k^3 & k^2 \\
k^3 & 0 & -k^1 \\
-k^2 & k^1 & 0
\end{pmatrix} K ij = − ϵ ijk k k = ⎝ ⎛ 0 k 3 − k 2 − k 3 0 k 1 k 2 − k 1 0 ⎠ ⎞ Note that ∣ k ∣ 2 = k 1 2 + k 2 2 + k 3 3 = 1 |\mathbf{k}|^2 = k_1^2 + k_2^2 + k_3^3 = 1 ∣ k ∣ 2 = k 1 2 + k 2 2 + k 3 3 = 1 . We get:
( K v ) i = K i j v j = − ϵ i j k k k v j = ϵ i k j k k v j = ( k × v ) i . (\mathbf{K}\mathbf{v})_i
=\mathbf{K}_{ij} v^j
= -\epsilon_{ijk} k^k v^j
= \epsilon_{ikj} k^k v^j
= (\mathbf{k}\times\mathbf{v})_i\,. ( Kv ) i = K ij v j = − ϵ ijk k k v j = ϵ ikj k k v j = ( k × v ) i . We will also need:
( K 2 ) i j = K i k K k j = ( − ϵ i k l k l ) ( − ϵ k j m k m ) = ( δ j l δ m i − δ j i δ m l ) k l k m = k i k j − δ i j , \left(\mathbf{K}^2\right)_{ij}
=K_i{}^k K_{kj}
=(-\epsilon_i{}^k{}_l\, k^l)(-\epsilon_{kjm} k^m)
=(\delta_{jl} \delta_{mi}-\delta_{ji}\delta_{ml}) k^l k^m
=k_i k_j - \delta_{ij}\,, ( K 2 ) ij = K i k K kj = ( − ϵ i k l k l ) ( − ϵ kjm k m ) = ( δ j l δ mi − δ ji δ m l ) k l k m = k i k j − δ ij , where we used δ m l k l k m = 1 \delta_{ml}k^l k^m=1 δ m l k l k m = 1 .
Now we compute the characteristic polynomial:
P ( t ) = det ( K − t I ) = − t 3 − t ( k 1 2 + k 2 2 + k 3 2 ) = − t 3 − t , P(t) = \det(\mathbf{K}-t\mathbf{I})=-t^3-t(k_1^2+k_2^2+k_3^2) = -t^3-t\,, P ( t ) = det ( K − t I ) = − t 3 − t ( k 1 2 + k 2 2 + k 3 2 ) = − t 3 − t , P ( K ) = 0 , P(\mathbf{K}) = 0\,, P ( K ) = 0 , − K 3 − K = 0 , -\mathbf{K}^3-\mathbf{K} = 0\,, − K 3 − K = 0 , K 3 = − K . \mathbf{K}^3=-\mathbf{K}\,. K 3 = − K . From the last equation we get K 4 = − K 2 \mathbf{K}^4=-\mathbf{K}^2 K 4 = − K 2 and
K 5 = − K 3 = K \mathbf{K}^5=-\mathbf{K}^3=\mathbf{K} K 5 = − K 3 = K and so on.
Then the rotation matrix R \mathbf{R} R is:
R = e θ K = ∑ k = 0 ∞ ( θ K ) k k ! = I + θ K + ( θ K ) 2 2 ! + ( θ K ) 3 3 ! + ⋯ = \mathbf{R}
= e^{\theta \mathbf{K}}
= \sum_{k=0}^\infty {(\theta\mathbf{K})^k \over k!}
= \mathbf{I} + \theta\mathbf{K}
+ {(\theta\mathbf{K})^2 \over 2!}
+ {(\theta\mathbf{K})^3 \over 3!}
+ \cdots = R = e θ K = k = 0 ∑ ∞ k ! ( θ K ) k = I + θ K + 2 ! ( θ K ) 2 + 3 ! ( θ K ) 3 + ⋯ = = I + K ( θ − θ 3 3 ! + θ 5 5 ! − ⋯ ) + K 2 ( θ 2 2 ! − θ 4 4 ! + θ 6 6 ! − ⋯ ) = = \mathbf{I}
+ \mathbf{K}\left(\theta - {\theta^3\over3!} + {\theta^5\over5!} - \cdots\right)
+ \mathbf{K}^2\left({\theta^2\over2!} - {\theta^4\over4!} + {\theta^6\over6!} - \cdots\right)
= = I + K ( θ − 3 ! θ 3 + 5 ! θ 5 − ⋯ ) + K 2 ( 2 ! θ 2 − 4 ! θ 4 + 6 ! θ 6 − ⋯ ) = = I + K sin θ + K 2 ( 1 − cos θ ) . = \mathbf{I}
+ \mathbf{K}\sin\theta
+ \mathbf{K}^2(1-\cos\theta)\,. = I + K sin θ + K 2 ( 1 − cos θ ) . So we got:
R = I + K sin θ + K 2 ( 1 − cos θ ) . \mathbf{R} = \mathbf{I}
+\mathbf{K} \sin\theta
+\mathbf{K}^2 (1-\cos\theta)\,. R = I + K sin θ + K 2 ( 1 − cos θ ) . Now we can write:
v r o t = R v \mathbf{v}_\mathrm{rot} = \mathbf{R} \mathbf{v} v rot = Rv and get:
v r o t = v + K v sin θ + K 2 v ( 1 − cos θ ) . \mathbf{v}_\mathrm{rot} = \mathbf{v}
+\mathbf{K}\mathbf{v} \sin\theta
+\mathbf{K}^2\mathbf{v} (1-\cos\theta)\,. v rot = v + Kv sin θ + K 2 v ( 1 − cos θ ) . We use:
K v = k × v , \mathbf{K}\mathbf{v} = \mathbf{k}\times\mathbf{v}\,, Kv = k × v , K 2 v = K ( K v ) = k × ( k × v ) = k ( k ⋅ v ) − ( k ⋅ k ) v = k ( k ⋅ v ) − v \mathbf{K}^2\mathbf{v} = \mathbf{K}(\mathbf{K}\mathbf{v})
= \mathbf{k}\times(\mathbf{k}\times\mathbf{v})
= \mathbf{k}(\mathbf{k}\cdot\mathbf{v}) - (\mathbf{k}\cdot\mathbf{k})\mathbf{v}
= \mathbf{k}(\mathbf{k}\cdot\mathbf{v}) - \mathbf{v} K 2 v = K ( Kv ) = k × ( k × v ) = k ( k ⋅ v ) − ( k ⋅ k ) v = k ( k ⋅ v ) − v and get:
v r o t = v cos θ + ( k × v ) sin θ + k ( k ⋅ v ) ( 1 − cos θ ) . \mathbf{v}_\mathrm{rot} = \mathbf{v} \cos\theta
+(\mathbf{k}\times\mathbf{v}) \sin\theta
+\mathbf{k}(\mathbf{k}\cdot\mathbf{v}) (1-\cos\theta)\,. v rot = v cos θ + ( k × v ) sin θ + k ( k ⋅ v ) ( 1 − cos θ ) . This is called the Rodrigues’ formula. In components:
R i j = I i j + K i j sin θ + ( K 2 ) i j ( 1 − cos θ ) . \mathbf{R}_{ij} = \mathbf{I}_{ij}
+\mathbf{K}_{ij} \sin\theta
+\left(\mathbf{K}^2\right)_{ij} (1-\cos\theta)\,. R ij = I ij + K ij sin θ + ( K 2 ) ij ( 1 − cos θ ) . We can simplify:
R i j = δ i j − ϵ i j k n k sin θ + ( n i n j − δ i j ) ( 1 − cos θ ) , R_{ij} = \delta_{ij}
-\epsilon_{ij}{}^k n_k \sin\theta
+(n_i n_j - \delta_{ij}) (1-\cos\theta)\,, R ij = δ ij − ϵ ij k n k sin θ + ( n i n j − δ ij ) ( 1 − cos θ ) , R i j = cos θ δ i j + ( 1 − cos θ ) n i n j − sin θ ϵ i j k n k . R_{ij}=\cos\theta\delta_{ij}+(1-\cos\theta)n_i n_j
-\sin\theta\,\epsilon_{ij}{}^k n_k\,. R ij = cos θ δ ij + ( 1 − cos θ ) n i n j − sin θ ϵ ij k n k .