Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Transformations and symmetries

Transformations of a quantum system

Above we see how tranlations and rotations act on a measurement process. Note that we apply them to the system itself (the red formless blob), the experimental apparatus (the box with the dial), the observer, and the environment (such as the gravitational field). We can think of these as active transformations, physically rotating everythingl or as passive transformations, corresponding to changes of coordinates.

Either way, the principle is that with such transformations, the results of experiments should be the same.

Transformations

Definition

To begin with, we give a precise definition of transformations acting on quantum systems. These are linear maps

T:ψψT:OO\begin{align} & {\cal T}: \ket{\psi} \to \ket{\psi'}\\ & {\cal T}: {\cal O} \to {\cal O}' \end{align}

for all states ψ\ket{\psi'} and operators O{\cal O}', such that observed quantities do not change. These are matrix elements:

ψχ2=ψχ2ψOψ=ψOψ\begin{align} |\brket{\psi'}{\chi'}|^2 & = |\brket{\psi}{\chi}|^2\\ \bra{\psi'}{\cal O}'\ket{\psi'} & = \bra{\psi}{\cal O}\ket{\psi} \end{align}

(you might think you want to know all matrix elements of O{\cal O}, but in practice measurements can be phrased as expectation values of projecttion operators as we have discussed -- also, it happens that if you know the expectation value of O{\cal O} for all states, you know all matrix elements. We’ll prove that at the end of this section so as not to interrupt the flow).

One can easily prove the following (see Bellac, 2006 or Messiah, 1999):

Theorem: give a transformation T{\cal T}, we can modify the transformed states ψeiα(ψ)ψ\ket{\psi'} \to e^{i\alpha(\ket{\psi'})}\ket{\psi'} such that the transformation is implemented by a linear operator UTU_{{\cal T}}:

ψ=UTψ ψO=UTOUT\begin{align} \ket{\psi'} & = U_{{\cal T}} \ket{\psi}\ \forall \ket{\psi}\\ {\cal O}' & = U_{{\cal T}} {\cal O} U_{{\cal T}}^{\dagger} \end{align}

such that UT=UT1U^{\dagger}_{{\cal T}} = U^{-1}_{{\cal T}} and UTU_{{\cal T}} is either:

  1. A linear operator (and thus unitary), or

  2. An antilinear operator, that is one satsifying

UT(aψ+bχ)=aUTψ+bUTχU_{{\cal T}}\left(a \ket{\psi} + b \ket{\chi}\right) = a^* U_{{\cal T}} \ket{\psi} + b^* U_{{\cal T}} \ket{\chi}

Antilinear operators of this kind are important in particle physics and quantum condensed matter physics -- in particular the transformation of time reversal is implemented by such an operator.

Before continuing, we’ll fulfill our promise by showing that knowledge of ψOψ\bra{\psi}\cO\ket{\psi} for all states ψ\ket{\psi} yields all matrix elements of O\cO. Let us say that we want to find αOβ\bra{\alpha} \cO \ket{\beta}. We define two linear combinations

χ1=α+βχ2=α+iβ\begin{align} \ket{\chi_1} & = \ket{\alpha} + \ket{\beta}\\ \ket{\chi_2} & = \ket{\alpha} + i \ket{\beta} \end{align}

Then

χ1Oχ1iχ2Oχ2=(1i)(αOα+βOβ)+2αOβ\bra{\chi_1} \cO \ket{\chi_1} - i \bra{\chi_2}\cO \ket{\chi_2} = (1 - i) \left(\bra{\alpha}\cO\ket{\alpha} + \bra{\beta}\cO\ket{\beta}\right) + 2 \bra{\alpha}\cO\ket{\beta}

That is, αOβ\bra{\alpha}\cO\ket{\beta} can be written in terms of expectation values of O\cO in various states.

Examples

It is worth mentioning a few examples:

  1. Translations. These are induced by the spatial transformation xx+a{\vec x} \to {\vec x} + {\vec a}. The action on wavefunctions would be ψ(x)ψ(xa)\psi({\vec x}) \to \psi({\vec x} - {\vec a}). As we’ve shown before, Ua=eip^a/U_{{\vec a}} = e^{- i {\hat{\vec p}}\cdot a/\hbar};

xUaψ=ψ(xa , \bra{{\vec x}}U_{{\vec a}}\ket{\psi} = \psi({\vec x} - {\vec a}\ ,

and

Uax^Ua^=x^+a U_{{\vec a}} {\hat{\vec x}} U_{{\hat a}}^{\dagger} = {\hat{\vec x}} + {\vec a}
  1. Rotations. These will be explained in great detail below.

  2. Parity. This is related to the spacetime transformation xx{\vec x} \to - {\vec x}. In quantum mechanics it is implemented by a unitary operator Π=Π=Π1\Pi = \Pi^{\dagger} = \Pi^{-1}. For states in L2(R3)L^2(\CR^3) we have

xΠψ=ψ(x)Πx^Π=x^Πp^Π=p^\begin{align} \bra{{\vec x}} \Pi \ket{\psi} & = \psi(-x) \\ \Pi {\hat{\vec x}} \Pi & = - {\hat{\vec x}}\\ \Pi {\hat{\vec p}} \Pi & = - {\hat{\vec p}} \end{align}
  1. Isospin. Nucleons have a spin degree of freedom and also an “isospin” degree of freedom; these contribute a factor H=C2C2\cH = \CC^2 \otimes \CC^2 to the hilbert space. Just as a good basis for the Hilbert space of a spin-12\half particle is the spin-up and spin-down states along some axis, a good basis for the isospin degree of freedom is the proton state and the neutron state. That these are degrees of freedom and not just labels of a particle is made clear by beta decay, in which a neutron can decay into a proton (by emitting a positron and an electron neutrino). This is a classic example of an internal symmetry, a symmetry that is not related to some transformation of spacetime.

  2. Time translations which take ψ(t)ψ(t+T)\ket{\psi(t)} \to \ket{\psi(t + T)}.

Transformation groups

Clearly a a sequence of transformations is itself a transformation. When implemented via linear operators, this arises because for U1,U2U_1, U_2 unitary operators, U1U2U_1 U_2 is also unitary as you can easily check. In fact, transformations form a group. It is worth stepping back and describing this mathematical object. Groups are a classic case of simple systems yielding rich structures; group theory yields powerful results and insights into quantum mechanics.

A group GG is a collection of objects g,h,...g, h, ... which could be countable (disceret) or uncountable, endoweed with a multiplication law such that ghGg\cdot h \in G \forall gm h \in G$. This multiplication law satisfies the following properties:

  1. Existence of an identity. There is an element eGe \in G such that eg=ge=ggGe g = g e = g \forall g \in G.

  2. Existence of an inverse. For every gGg \in G, there exists a group element g1Gg^{-1} \in G such that gg1=g1g=eg g^{-1} = g^{-1} g = e.

  3. Associativity. For any f,g,gGf, g, g \in G,

f(gh)=(fg)hf \cdot (g \cdot h) = (f \cdot g) \cdot h

where one does the multiplication in parentheses first.

Note that in general it is not true that ghhgg\cdot h h \cdot g. Groups that satisfy this property are called Abelian; otherwise they are called non-Abelian.

There are many examples of groups. A few are

  1. Discrete translations in space xx+iniai{\vec x} \to {\vec x} + \sum_i n_i {\vec a}^i, for a{\vec a} a collection of linearly independent vectors and niZn_i \in \CZ. This is clearly an Abelian group; “multiplication” in this case is simply vector addition.

  2. Permutations. These are one-to-one and onto transformations of nn elements into themselves:

σ:(1,2,n)=(σ(1),σ(2),,σ(n))\sigma : (1,2,\ldots n) = (\sigma(1), \sigma(2), \ldots, \sigma(n))

This group is often called SnS_n, and it is non-Abelian if n>2n > 2; I recommend convincing yourself of this by playing with S3S_3.

  1. n×nn \times n invertible matrices (GL(n,R)GL(n, \CR) for real matrices or GL(n,C)GL(n,\CC) for complex matrices). These form a non-Abelian group under matrix multiplication.

Unitary representations

For a group GG, a unitary representation is a Hilbert space H\cH together with a map gUgg \to U_g of group elements to unitary operators that respects the group transformation law

Ug1Ug2=Ug1g2U_{g_1} U_{g_2} = U_{g_1\cdot g_2}

That is to say the map from G to unitary operators is a group homomorphism. Note that this reuires that Ue=1U_e = {\bf 1}, Ug1=UgU_{g^{-1}} = U_g^{\dagger}.

As a side note: there are also projective representations, which correspond to the above map but a modified multiplication law:

Ug1Ug2=eiα(g1,g2)Ug1g2U_{g_1} U_{g_2} = e^{i\alpha(g_1,g_2)} U_{g_1\cdot g_2}

where α\alpha is a real number.

An important concept is that of a {\it irreducible representation}, often called {\it irreps}\ from which all representations can be built. The theory of group representations is all about the properties of the irreps.

To start with, give a representation H\cH of a group GG, a subspace HH\cH' \subset \cH is an {\it invariant subspace} if for energy ψH\ket{\psi'} \in \cH', and gGg \in G, UgψHU_{g} \ket{\psi'} \in \cH'. An example for real representations would be that the xyx-y plane is invariant under the group of rotations about the zz axis. For that matter, the zz axis is also an invariant subspace!

Next, a representation H\cH of GG is {\it irreducible}\ if the obly invariant subspace under actions of GG is H\cH itself. Otherwise, H\cH is {\it reducible}. Furothermore, for a reducible representation H\cH, we can always write

H=H1H2Hn\cH = \cH_1 \oplus \cH_2 \oplus \cdots \oplus \cH_n

where Hk=1,,n\cH_{k = 1,\ldots, n} are all irreducible. These factors will all be orthogonal in the Hilbert space.

Symmetries

A particularly important class of transformations are symmetries: transformations that preserve the system’s dynamics. More precisely, they preserve the Hamiltonian:

H=UTHUT=H[H,UT]=0H' = U_{{\cal T}} H U_{{\cal T}}^{\dagger} = H \Rightarrow [H,U_{{\cal T}}] = 0

The essential point is that eigenstates of UU will also be eigenstates of HH (this will become more obvious, perhaps, when we consider infinitesimal transformations, for which UU can be approximated by a Hermitian operator). Thus, for example, H=p22mH = \frac{p^2}{2m} is a

In practice we are often interested in approximate symmetries. This means that we can write

H=H0+ΔH = H_0 + \Delta

such that UTH0UT=H0U_{{\cal T}} H_0 U_{{\cal T}}^{\dagger} = H_0 but UTΔUTΔU_{{\cal T}} \Delta U_{{\cal T}}^{\dagger} \neq \Delta. This is an “approximate symmetry” if ψH0ψψΔψ\bra{\psi}H_0\ket{\psi} \gg \bra{\psi}\Delta\ket{\psi} for states ψ\psi of interest (that is, for states which appear in measured processes). An important example is nuclear isospin, which transforms protons into neutrons. If this was an exact symmetry, this transformation would preserve the eigenstate of the Hamiltonian, which would include the rest mass of said particles. But of course in practice, the rest masses are not the same, though they are close: mpc2=938.272MeVm_pc^2 = 938.272 MeV and mnc2=939.565MeVm_n c^2 = 939.565 MeV. Roughly we can write

H=Hstrong+Hem+HweakH = H_{strong} + H_{em} + H_{weak}

for the interaction between nucleaons and other fundamental particles. At internucleon distances, the strong nuclear force dominates, and preserves nuclear isospin, The electromagnetic and weak interactions break it (for example via coulomb repulsion of protons as opposed to neutrons), but it is weak at these scales.

Symmetries and irreps

Given a group GG of transformations, the Hilbert space of our quantum system generically forms a reducible representation, which can be written as a direct sum of irreps.

If the group GG is a symmetry, then any degenerate eigenspace of HH is a representation of the symmetry group. This is because any transformation by the group GG will not change the energy since GG commutes with the Hamiltonian. By the statement above, this representation can thus be written as a direct sum of irreps. We can do this for every energy eigenstate in the Hilbert space, this writing the Hilbert space as a direct sum of irreps, each irrep living in a degenerate subspace of HH. From this we see that every irrep of a symmetry group GG is an eigenstace of HH (it could be one-dimensional or higher-dimensional).

Continous vs. Discrete Transformations

  1. Discrete transformations correspond to groups which have a countable set of elements. Examples are:

  1. Continuous transformations. These are transformations labelled by families of continuous real parameters t1,,tNt_1,\ldots,t_N. NN here would be the dimension of the group -- the number of independent coordinates needed to describe its elements. (We could also consider complex parameters of course; still these can be written in terms of their real and imaginary parts so we will stick to real parameters)

Examples:

Infinitesimal transformations

Now let us consider a continuous family of transformations that includes the identity. We choose coordinates such that the identity is the origin of the coordinate system.

By infinitesimal tranformations we mean ones for which UU (or more precisely, its matrix elements) can be well approximated by an O(tk)\cO(t_k) approximation, that is

U1itkAk+O(t2)U \sim {\bf 1} - i t_k A_k + \cO(t^2)

where AkA_k is some operator. The demand that UU be unitary at O(t)\cO(t) means:

1=UU(1+itkAk)(1itkAk)+O(t2)=1+itk(AkAk)+O(t2){\bf 1} = U U^{\dagger} \sim ({\bf 1} + i t_k A_k) ({\bf 1} - i t_k A^{\dagger}_k) + \cO(t^2) = 1 + i t_k(A_k - A_k^{\dagger}) + \cO(t^2)

In other words, we must have Ak=AkA_k = A_k^{\dagger}: infinitesimal transformations are described by Hermitian operators.

Let us write $t_k A_k From an infinitesimal transformation. We can build up a finite transformations via

U(t)=eitAU(t) = e^{i t A}

It is easy to show that the set of operators U(t)U(t) form an Abelian group (a subgroup of the transformation group) with U(t1)U(t2)=U(t1+t2)U(t_1) U(t_2) = U(t_1 + t_2), for which U(t)=UU(-t) = U^{\dagger}. Depending on the eigenvalues of AA, this group is equivalent to either R\CR or S1S^1; the latter will appear if the eigenvalues of AA are integer multiples of some basic interval δt\delta t, so that ttδtt \equiv t _ \delta t. We say that the Hermitian operators generate the symmetry.

If we have two unitary operators of the form UA=eitAU_A = e^{i t A} and UB=eisBU_B = e^{i s B}, the lack of commutativity of the unitary operators can be related to the lack of commutativity of A,BA,B. To see this, compute

UB1UA1UBUA=(1isB12s2B2)((1itA12t2A2)(1+isB12s2B2)(1+itA12t2A2)+=1+st(ABBA)+\begin{align} U_B^{-1} U_A^{-1} U_B U_A & = (1 - i s B - \half s^2 B^2)( (1 - i t A - \half t^2 A^2) (1 + i s B - \half s^2 B^2) (1 + i t A - \half t^2 A^2) + \ldots & = 1 + st (AB - BA) + \ldots \end{align}

So the unitary operators commue if their infinitesimal generators commute.

Infinitesimal symmetries

Demanding U(t)HU(t)=HU(t) H U(t)^{\dagger} = H for an infinitesimal symmetry means

(1+itA+)H(1itA)=Hit[A,H]+=H(1 + i t A + \ldots) H (1 - i t A \ldots) = H _ i t [A,H] + \ldots = H

which implies [A,H]=0[A,H] = 0. This is called an infinitesimal symmetry. The generator is a Hermitian operator which commutes with the Hamiltonian. This, it is associated to a conserved quantity. For example, you can show that

A=ψ(t)Aψ(t)\vev{A} = \bra{\psi(t)} A \ket{\psi(t)}

is independent of time, since

ddtA=1i[A,H]=0\frac{d}{dt} \vev{A} = \frac{1}{i\hbar}\vev{[A,H]} = 0

Similarly, if Aψ(t)=aψ(t)A\ket{\psi(t)} = a \ket{\psi(t)} at some time tt, with aa an eigenvalue of AA, this equation will be true for all tt. This is the quantum analog of Noether’s theorem.

Two important examples:

  1. Time translations. If HH is time-independent, then since [H,H]=0[H, H] = 0 and U(t)=eitH/U(t) = e^{- i t H/\hbar} is the unitary operator generating time translations, time translations is a symetry of the theory and is tied to energy conservation.

  2. Spatial translations. Infinitesimal transformations are generated by the momentum operator. If H=H(p)H = H({\vec p}), then [p,H]=0[p, H] = 0 and momentum is conserved.

References
  1. Bellac, M. L. (2006). Quantum Physics. Cambridge University Press. https://books.google.com/books?id=uSQ-jwEACAAJ
  2. Messiah, A. (1999). Quantum Mechanics. Dover Publications. https://books.google.com/books?id=mwssSDXzkNcC