Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Entangled states and density matrices

Direct sums and tensor products

There are different ways to combine vector spaces to get new vectors spaces. We are particularly interested in the tensor product, but the direct sum is also important and the contrast is illuminating

Direct sums

The direct sum V=V1V2V = V_1 \oplus V_2 is a method for adding new directions to the Hilbert space. For example, R3\CR^3 can be thought of as a combination of R2\CR^2 represented as the xyx-y plane, and R\CR thought of as the zz axis. Another example would be a particle hopping between two sites; in one site there are three states. If these sites were ions and the particle an electron, they could be possible excitation levels of the particle at that site. The particle could be at either one site or the other, in any state associated to that site, or a linear combination of the two sites.

c(v1,v2)+d(w1,w2)=(cv1+dw1,cv2,dw2)c \left(\ket{v}_1, \ket{v}_2\right) + d \left(\ket{w}_1, \ket{w}_2\right) = \left(c\ket{v}_1 + d \ket{w}_1, c \ket{v}_2, d \ket{w}_2\right)

for all c,dCc,d \in \CC. If we define the zero vector as (0,0)\left(0,0\right), this is a vector space.

Typically, we drop the ordered paits and simply write

(v1,v2)=v1+v2\left(\ket{v}_1,\ket{v}_2\right) = \ket{v}_1 + \ket{v_2}

which can be understood formally if we understand v1(v1,0)\ket{v}_1 \equiv \left(\ket{v}_1, 0\right) and v2=(0,v2)\ket{v}_2 = \left(0,\ket{v}_2\right).

v1=(c1c2) ;  v2=(d1d2d3) ;  (v1,v2)=(c1c2d1d2d3)\ket{v}_1 = \begin{pmatrix} c_1 \\ c_2 \end{pmatrix}\ ; \ \ \ket{v}_2 = \begin{pmatrix} d_1 \\ d_2 \\ d_3 \end{pmatrix}\ ; \ \ \left(\ket{v}_1, \ket{v_2}\right) = \begin{pmatrix} c_1 \\ c_2 \\ d_1 \\ d_2 \\ d_3 \end{pmatrix}
MM5(C)=(A2×211A2×312A3×221A3×322)M \in M_5(\CC) = \begin{pmatrix} A^{11}_{2\times 2} & A^{12}_{2\times 3} \\ A^{21}_{3\times 2} & A^{22}_{3\times 3} \end{pmatrix}

then A11,A22A^{11}, A^{22} are the embeddings of linear operators acting on V1,V2V_1,V_2 respectively, while A12,A210A^{12}, A^{21}\neq 0 lead to new operators that only have meaning in the direct sum.

As an example, if V1V_1 corresponds to the states of an electron on one site, and V2V_2 the states of an electron on another site, then of the Hamiltonian is a sum H=H1+H2H = H_1 + H_2 with HiH_i acting on ViV_i, the eigenstates correspond to particles living on one site or another, and if the particle starts on a site it will stay on that site. On the other hand, if HH contains terms that mix the two sites, the eigenstates are lienar combinations of the particle being on one site or another; relatedly, the state vector of a partcile that starts on one site will acquire a component correspond to the particle siting on the other site.

Tensor products

The tensor product emerges when you want to add an attribute to a given system. That is, consider a particle for which you can specify the position and spin degree of freedom. In this case the particle has position and spin, you can’t either be in position x{\vec x} or have spin-up along the zz direction. Similarly, you might want to describe the states of two distinguishable particles such as a proton and an electron. You always have to say what each particle is doing.

The definition is a little more formal.

α=i,IαiIiketI\ket{\alpha} = \sum_{i, I} \alpha_{iI} \ket{i}ket{I}

If β\ket{\beta} is similarly defined with coefficients βiI\beta_{iI}, then addition is defined as

cα+dβ=i,I(cαiI+dβiI)iIc \ket{\alpha} + d \ket{\beta} = \sum_{i,I} (c \alpha_{iI} + d \beta_{iI})\ket{i}\ket{I}

for any c,dCc,d \in \CC. If we define the zero vector via αiI=0\alpha_{iI} = 0, then V1V2V_1\otimes V_2 is a vector space.

v1w2=i,IviWIiI\ket{v}_1 \ket{w}_2 = \sum_{i,I} v_i W_I \ket{i}\ket{I}
α=+,zψ++,zψ\ket{\alpha} = \ket{+,z}\ket{\psi_+} + \ket{-,z} \ket{\psi_-}

where ψ±\psi_{\pm} denote square-integrable functions. This is sometimes realized as a spinor, a 2d vector of wavefunctions

α(ψ+(x)ψ(x))\ket{\alpha} \Rightarrow \begin{pmatrix} \psi_+({\vec x}) \\ \psi_-({\vec x}) \end{pmatrix}
A1A2vw=(A1v)(A2w)A_1 \otimes A_2 \ket{v}\ket{w} = (A_1\ket{v})(A_2 \ket{w})

We can further add operators such as A1A2+B1B2A_1 \otimes A_2 + B_1 \otimes B_2 so that

(A1A2+B1B2)vw=(A1v)(A2w)+(B1v)(B2w)(A_1 \otimes A_2 + B_1 \otimes B_2) \ket{v}\ket{w} = (A_1\ket{v})(A_2 \ket{w}) + (B_1\ket{v})(B_2 \ket{w})
βα=w1v1w2v2\brket{\beta}{\alpha} = \brket{w_1}{v_1} \brket{w_2}{v_2}
H=H11+1H2+gkAkBkH = H_1 \otimes {\bf 1} + {\bf 1}\otimes H_2 + g \sum_k A_k \otimes B_k

where gg is a numnber included for convenience. When g=0g = 0, the factors of teh tensor product are non-interacting. For the 2-particle case, what one particle is doing does not affect what the other paerticle is doing. However, if g0g \neq 0, the particles interact; the state of one affects the dynamics of the state of the other. Imagine two spin-12\half particles. The Hamkiltonian

H=μ1BS1,z1μ2B1S2,z+giS1,iS2,iH = - \mu_1 B S_{1,z} \otimes {\bf 1} - \mu_2 B {\bf 1}\otimes S_{2,z} + g \sum_{i}S_{1,i} \otimes S_{2,i}

can be thought of as two magnetic momnents in a magnetic filed B=Bz^{\vec B} = B {\hat z}. The last term can be thought of as saying that each particle feels a magnetic field proportional to the magnetic moment of the other particle.

Entangled states

Given a tensor product Hilbert space H=H1H2\cH = \cH_1 \otimes \cH_2, we say that a state αH\ket{\alpha} \in \cH is entangled if it cannot be written in the form

α=v1w2\ket{\alpha} = \ket{v}_1 \ket{w}_2

An example works best to illustrate this. Consider two spin-12\half particles. The following states form a basis of C2C2\CC^2\otimes\CC^2:

ψ1=+,z+,zψ2=12(+,z,z+,z+,z)ψ3=,z,zψ4=12(+,z,z,z+,z)\begin{align} \ket{\psi_1} & = \ket{+,z}\ket{+,z}\\ \ket{\psi_2} & = \frac{1}{\sqrt{2}}\left(\ket{+,z}\ket{-,z} + \ket{-,z}\ket{+,z}\right)\\ \ket{\psi_3} & = \ket{-,z}\ket{-,z}\\ \ket{\psi_4} & = \frac{1}{\sqrt{2}} \left(\ket{+,z}\ket{-,z} - \ket{-,z}\ket{+,z}\right) \end{align}

ψ1,2\ket{\psi_{1,2}} are clearly unentangled. ψ2,4\ket{\psi_{2,4}} are entangled and in fact there is no way to write them as unentangled states. We will show a tool that can test whether the state of two systems is entangled or not, when we introduce density matrices and the von Neumann entropy.

As we can see from the above, quantum entanglement denotes a correlation between the state of two degrees of freedom in a tensor product. This yields some surprising results. For example, assume that we prepare the two spin states of a particle in ψ4\ket{\psi_4}. We can do this in such a way that the two particles are separated by a large distance. Even at this separation, the results of measuring SzS_z for the first particle are always correlated with the results of measuring SzS_z for the second particle: if one is measured as spin up, the other is measured as spin down. This is true even if the measurements are made simultaneously, so that no light signal could pass between the two particles. This is the famous Einstein-Podolsky-Rosen paradox. Note that since the particles would have to be prepared with the spins so entangled when they were in causal contact, and then separated -- there is nothing acausal about how the states were constructed. The puzzle is over whether there is anything acausal about the measurement process. Note that there is no way to use EPR states to communicate, since an experimentalist measuring SzS_z for either particle will get each of ±2\pm \frac{\hbar}{2} 50%50\% of the time; they have no control over the result, and only know it is correlated with the spin of the other particle if they knew in advance the pair were so prepared. If causality in a special-relativistic context means simply ``information cannot be transmitted at speeds faster than light" then it is arguably safe here.

Density matrices

Density matrices capture cases where we have imperfect kmowledge of the quantum system. They typically arise because the state of the system we are interested in studying was prepared through coupling to degrees of freedom whose state we do not observe, and the system’s state is entangled with the unobserved degrees of freedom.

We will thuis start with a purely quantum description of such a stuation, in which both the observed “system” is coupled to some unobserved degrees of freedom, called the “environment”. This setup is a subject of intensive study which goes onder the name of {\it open quantum systems}. We will then provide a more phenomenological output in which the density matrix reflects some classical uncertainty regarding the preparation of the state.

Open quantum systems.

In open quantum systems we imagine that the Hilbert spacecan be written as a tensor product H=HsysHenv\cH = \cH_{sys}\otimes \cH_{env}. Some examples:

  1. {\it Local region of an extended system}.

Spin chain

Pictured above is a one-dimensional spin chain; a lattice of sites with each site oddupied by a spin-12\half particle. If theer are NN sites, the Hilbert space is C2C2C2=(C2)NC2N\CC^2 \otimes \CC ^2\otimes \ldots \CC^2 = (\CC^2)^{\otimes N} \cong \CC^{2N}; that is, NN copies of C2\CC^2 combined in a tensor product structure. We can imagine that the observed spins correspond to M<NM < N contiguous sites (in the picture, M=6M = 6).

  1. {\it Particle in imperfect vacuum}. Here Hsys\cH_{sys} is the Hilbert space of some particle; Henv\cH_{env} is the Hilbert space of soem unobserved gas particles which can interact/collide with the observed particle.

If i\ket{i} is an othonormal basis for Hsys\cH_{sys} and I\ket{I} an orthonormalbasis for Henv\cH_{env}, then as we discussed above, a general state of H\cH is:

ψ=iIαiIiI\ket{\psi} = \sum_{iI} \alpha_{iI} \ket{i}\ket{I}

We will assume that ψψ=1\brket{\psi}{\psi} = 1.

Now let us say we can only measure the system. The corresponding operator ha sthe following structure in H\cH:

A=Asys1A = A_{sys}\otimes {\bf 1}

A natural question to ask is the expectation value of AA, which is:

A=iIIiαiIjJαjJ(AsysjJ=iI,jJαiIαjJiAsysjIJ=i,j,IαiIαjIiAsysj=i,j,k,IδjkαiIαjIiAsysj=i,j,k,IkjαiIαjIiAsysk=ki,j,IkjαiIαjIiAsysk\begin{align} \vev{A} & = \sum_{iI} \bra{I}\bra{i} \alpha^*_{iI} \sum_{jJ} \alpha_{jJ} (A_{sys}\ket{j}\ket{J}\\ & = \sum_{iI,jJ} \alpha^*_{iI}\alpha_{jJ} \bra{i} A_{sys}\bra{j} \brket{I}{J}\\ & = \sum_{i,j,I} \alpha^*_{iI}\alpha_{jI} \bra{i} A_{sys}\bra{j}\\ & = \sum_{i,j,k,I}\delta_{jk} \alpha^*_{iI}\alpha_{jI} \bra{i} A_{sys}\ket{j}\\ & = \sum_{i,j,k,I} \brket{k}{j} \alpha^*_{iI}\alpha_{jI} \bra{i} A_{sys}\ket{k}\\ & = \sum_k \sum_{i,j,I}\brket{k}{j} \alpha^*_{iI}\alpha_{jI} \bra{i}A_{sys}\ket{k}\\ \end{align}

Now define

ρ=ijIjαjIαiIi\rho = \sum_{ijI} \ket{j} \alpha_{jI} \alpha^*_{iI} \bra{i}

Since for any linear operator OO acting on Hsys\cH_{sys}, trO=kkOk\text{tr} O = \sum_k \bra{k} O \ket{k}, we have

A=trρA\vev{A} = \text{tr} \rho A

Recall that from the rules of quantum mechanics, we can express the probability of geting a specific value aa of AsysA_{sys} as the expectation value of the projection operator Pa\CP_a. Thus the above rule for expectation values yields:

p(a)=tr(ρPa)p(a) = \text{tr}(\rho \CP_a)

Before continuing, I wil note that it is useful to think of the density matrix constructed above as a “partial trace” over the environmental degrees of freedom:

trHenvψψ=IIψψI=I,jJ,kKαjJIJjαkKkKI=jkIjαjIαkIk=ρ\begin{align} \text{tr}_{\cH_{env}}\ket{\psi}\bra{\psi} & = \sum_{I} \brket{I}{\psi}\brket{\psi}{I}\\ & = \sum_{I,jJ,kK}\alpha_{jJ} \brket{I}{J} \ket{j} \alpha^*_{kK} \bra{k} \brket{K}{I}\\ & = \sum_{jkI} \ket{j} \alpha_{jI}\alpha^*_{kI}\ket{k}\\ & = \rho \end{align}

This language is used frequently in the quantum information literature.

Phenomenological definition

The second starting point is to assume that the quantum system is in one of an ensemble of states governed by a classical probability distribution. You could imagine somebody preparing a spin state as +,z\ket{+,z} or ,x\ket{-,x} each with 50%50\% probability, for example. Note that this probability is distinct from the probabilty of a measurement of a known observable in a known quantum state yielding a particular outcome.

In the case at hand, if the state is prepared a large number of times, the average value of the measurement of some observable AA will be

A=ψpψψAψ\vev{A} = \sum_{\ket{\psi}} p_{\ket{\psi}}\bra{\psi} A \ket{\psi}

By a similar calculation as above, we can define

ρ=ψpψψψ\rho = \sum_{\psi} p_{\psi} \ket{\psi}\bra{\psi}

where the states ψ\ket{\psi} have unit norm, and pψp_{\psi} is the probability that the system was prepared in state ψ\ket{\psi}. Note that we have made no demand that the collection of states ψ\ket{\psi} are orthogonal to each other. We simply demand that ψ\ket{\psi} has unit norm, that 0pψ10 \leq p_{\ket{\psi}} \leq 1, and that ψpψ=1\sum_{\ket{\psi}} p_{\ket{\psi}} = 1.

We call the collection of states ψ\ket{\psi} together with the probabilities [ψ[_{\psi} an ensemble. A particularly important one is the canonical ensemble. If the system is not dynamically coupled to the environment, HH is the Hamiltonian of the system and E,α\ket{E,\alpha} is an orthonormal basis of eigenstates such that H]ketE,α=EE,αH]ket{E,\alpha} = E \ket{E,\alpha}, we define

pE,α=eβEZ(β)p_{E,\alpha} = \frac{e^{-\beta E}}{Z(\beta)}

where β=1/kBT\beta = 1/k_B T is the inverse temperature, and Z=E,αeβEZ = \sum_{E,\alpha} e^{-\beta E} is the partition function. This ensemble is the basis of quantum equilibrium statistical mechanics. The density matrix is then

ρβ=E,αeβEE,αE,αZ(β)\rho_{\beta} = \frac{\sum_{E,\alpha} e^{-\beta E} \ket{E,\alpha} \bra{E,\alpha}}{Z(\beta)}

The expectation value of an aobservable AA in this case is

Aβ=E,αeβEE,αAE,αZ(β)\vev{A}_{\beta} = \frac{\sum_{E,\alpha} e^{-\beta E} \bra{E,\alpha} A \ket{E,\alpha}}{Z(\beta)}

General properties of the density matrix

The following properties can be deduced from either of the above definitions.

  1. ρ\rho is Hermitian. Note that this means an orthonormal basis μ\ket{\mu} exists for which

ρ=μpμμμ\rho = \sum_{\mu} p_{\mu} \ket{\mu}\bra{\mu}

Inb the second of teh above construuctions of ρ\rho, μ\ket{\mu} is not necessarily the collection ψ\ket{\psi}; for starters, ψ\ket{\psi} do not have to be orthogonal.

  1. trρ=1\text{tr}\rho = 1. In the first case, this follows from the demandthat ψψ=1\brket{\psi}{\psi} = 1 for thestate of the combined system. In the second case, it follows from the demand that ψψ=1\brket{\psi}{\psi} = 1 for every state in the classical ensemble, and that ψpψ=1\sum_{\ket{\psi}} p_{\ket{\psi}} = 1.

  2. φρφ0 φHsys\bra{\varphi} \rho \ket{\varphi} \geq 0 \forall\ \ket{\varphi} \in \cH_{sys}. This implies that pμ0p_{\mu} \geq 0.

With these definitions, pμp_{\mu} can also be interpreted in terms of probabilities. The formula for the expectation value of some operator becomes

A=tr(ρA)=μpμμAμ\vev{A} = \text{tr}(\rho A) = \sum_{\mu} p_{\mu} \bra{\mu} A \ket{\mu}

Time evolution of the density operator

In general the system and the environment will be dynamically coupled. Energy can ass betwene the system and the environment. If HH is the Hamiltonian actig on the full Hilbert space HsysHenv\cH_{sys}\otimes\cH_{env}, then using the time-dependent Schroedinger equation for the full system, we find with some algebra that

iρt=(HiI;jJcjJckIHjJ;kIciIcjJ)iki\hbar \frac{\del \rho}{\del t} = (H_{iI;jJ}c_{jJ}c^*_{kI} - H_{jJ;kI} c_{iI}c^*_{jJ}) \ket{i}\bra{k}

This is generally a complicated “master equation” for ρ\rho. In certain specific situations it can take a much more specific form.

When the system and environment are decoupled after the full state has been prepared; or if the density matrix is contructed from a classical probability ensemble which is time-independent, you can show easily that

iρt=[Hsys,ρ]i\hbar \frac{\del \rho}{\del t} = [H_{sys},\rho]

This follows from the time-dependent Schroedinger equation itψ=Hsysψi\hbar \del_t \ket{\psi} = H_{sys}\ket{\psi}, together with its conjugate

itψ=ψH-i\hbar \del_t \bra{\psi} = \bra{\psi} H

This latter case is the one described in Commins, and in most introductory textbooks on the subject.

Pure vs. Mixed States and the von Neumann entropy

By convention, the state of the system is specified by the density operator. Our previous usage in terms of vectors in a Hilbert space is equivalent to the following:

You can show that the rule A=tr(ρA)=ψAψ\vev{A} = \text{tr}(\rho A) = \bra{\psi}A\ket{\psi} gives you everything you need to recover the results of measurements and so on.

SvN=trρlnρS_{vN} = - \text{tr} \rho \ln \rho

Taking the log of an operator is not trivial in general. If we wrote A=1+(1A)A = 1 + (1 - A), we can write it as a power series in (1A)(1-A) using ln(1+x)=1+x+12x2+\ln (1 + x) = 1 + x + \half x^2 + \ldots. There are techniques for computing SvNS_{vN} and it is a big industry in modern quantum theory. For now we will note that if we can write ρ\rho in an orthonormal basis, ρ=kpkkk\rho = \sum_k p_k \ket{k}\bra{k} with kl=δkl\brket{k}{l} = \delta_{kl}, then

SvN=kpklnpkS_{vN} = - \sum_k p_k \ln p_k

Recall that 0pk10\leq p_k \leq 1, and kpk=1\sum_k p_k = 1. Those familiar with classical information theory will recognize this as the Shannon entropy for a probability ensemble pkp_k where kk is treated as a random variable.

There is a big industry in computing the Shannon entropy, particular for subsystems of entangled states, as this is a measure of the degree of entanglement.