Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Linear operators

In classical mechanics we saw that the states of the system were described as points on phase space. Observables of the system were functions on phase space, that could be described as generators of infinitesimal transformations.

In quantum mechanics, the space of ppossible states are described as vectors in a complex vector space (recall the discussion of photon polarization). Observables and generators of transformations are described as operators which map vectors to other vectors. More specifically, the correspond to linear operators which respect the linear structure of the vector space (addition and scalkar multiplication).

Definitions

  1. Let V1,V2V_1,V_2 be two vector spaces over F\CF. A linear map FF is a map

f:V1V2f: V_1 \to V_2

such that

f(av+bw)=af(v)+bf(w)   a,bF ,  v,wV1f(a \ket{v} + b \ket{w}) = a f(\ket{v}) + b f(\ket{w})\ \ \ \forall a,b \in \CF\ ,\ \ \ket{v},\ket{w} \in V_1

Note that we can easily prove with this that for the zero vector 01V1\ket{0}_1 \in V_1, f(01)=02V2f(\ket{0}_1) = \ket{0}_2 \in V_2 where 02\ket{0}_2 is the zero vector.

  1. For vector spaces V1,V2V_1, V_2, over the complex numbers C\CC, an antilinear map ff is a map f:V1V2f: V_1 \to V_2 such that

f(av+bw)=af(v)+bf(w)   a,b F ,  v,wV1f(a \ket{v} + b \ket{w}) = a^* f(\ket{v}) + b^* f(\ket{w})\ \ \ \forall a,b \in \ \CF\ ,\ \ \ket{v},\ket{w} \in V_1

where a,ba^*, b^* are the complex conjugates of a,ba,b respectively. This is important for defining adjoint vectors and normas of vectors.

  1. Let VV be a vector space over F\CF. A linear operator f is a linear map F:VVF: V \to V.

For F=C\CF = \CC, an antilinear operator g is an antilinear map g:VVg: V \to V.

Antilinear operators include the operator that implements time reversal, which is important in various condensed matter physics and particle physics contexts.

Abstractly we usually denote operators by capital letters AA and the action of operators as A:VAvA: \ket{V} \to A\ket{v}.

Examples

  1. The identity operator 1v=v  v{\bf 1}\ket{v} = \ket{v}\ \forall\ \ket{v}.

  2. The zero operator :V0\varnothing: \ket{V} \to \ket{0}.

  3. More generally, scalar multiplication by any scalar in F\CF is a linear operator.

  4. For C=CnC = \CC^n, n×nn \times n matrices acting on column vectors by the usual rules of matrix multiplication are all linear operators. In fact one can represent any finite-dimensional vector space of dimension nn by Cn\CC^n, and every operator as a matrix acting on Cn\CC^n.

  5. A more interesting example: consider the (infinite-dimensional) space of all polynomials. We can represent these as

a0,a1,a2,=k=0akxk\ket{a_0,a_1,a_2,\ldots} = \sum_{k = 0}^{\infty} a_k x^k

The following operators

Sa0,a1,=0,a0,a1,=k=0akxk+1=xk=0akxkS\ket{a_0,a_1,\ldots} = \ket{0,a_0,a_1,\ldots} = \sum_{k = 0}^{\infty} a_k x^{k+1} = x \sum_{k = 0}^{\infty} a_k x^{k}
Ta0,a1,=a1,2a2,3a3,=k=1kk1kxk=ddxk=0akxkT \ket{a_0,a_1,\ldots} = \ket{a_1,2 a_2,3 a_3,\ldots} = \sum_{k = 1}^{\infty}k_{k-1} k x^k = \frac{d}{dx} \sum_{k = 0}^{\infty} a_k x^k

are linear operators on VV.

  1. Relatedly, consider the vector space of square-integrable complex functions. You can convince yourself that multiplication by xx and differentiation are linear operators. There are some subtleties here; it can be that these take the function out of the Hilbert space (eg multiplication can, if ff falls off as 1/x1/|x| at infinity).

Operator algebras

In general we can combine operators to form new linear operator in various ways: technically, they define an algebra:

  1. Operators can be added: for operators A1,A2A_1,A_2 acting on a vector space VV, we can define

(aA1+bA2)va(A1v)+b(A2v)(a A_1 + b A_2)\ket{v} \equiv a (A_1 \ket{v}) + b (A_2 \ket{v})

for any vV\ket{v} \in V, and a,bFa,b \in \CF.

It is straightforward to see that (aA1+bA2)(a A_1 + b A_2) is thus a linear operator, Note that A+=AA + \varnothing = A under this rule, A1+A2=A2+A1A_1 + A_2 = A_2 + A_1, and if a=0a = 0, aA=a A = \varnothing.

  1. We can also multiply operators. For any two linear operators A1,A2A_1,A_2, we define

(A1A2)v=A1(A2v)(A_1 A_2)\ket{v} = A_1(A_2\ket{v})

With a little work you can show that A1A2A_1 A_2 is a linear operator. Note that if 1\bf{1} is the identity operator, 1A=A1=A\bf{1} A = A \bf{1} = A. Multiplication is associative: that is, if A3A_3 is also a linear operator, ((A1A2)A3)=(A1(A2A3))((A_1 A_2) A_3) = (A_1(A_2 A_3)). However, it does not have to be commutative. For example, for V=CnV = \CC^n with operators equal to n×nn\times n matrices, operator multiplication is just matrix multiplication which is certainly not commutative. Consider V=C2V = \CC^2, and operators

A1=(0110) ;  A2=(1001) A_1 = \begin{pmatrix} 0 & 1 \\ 1 & 0\end{pmatrix}\ ; \ \ A_2 = \begin{pmatrix} 1 & 0 \\ 0 & -1 \end{pmatrix}

then operator multiplication is simply matrix multiplication which we know is not commutative. In part

A1A2=(0110) ;  A2A1=(0110)=A1A2A1A2A_1 A_2 = \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix}\ ; \ \ A_2 A_1 = \begin{pmatrix} 0 & 1 \\ -1 & 0 \end{pmatrix} = - A_1 A_2 \neq A_1 A_2

In general we measure the lack of commutativity with the commutator:

[A1,A2]=A1A2A2A1[A_1,A_2] = A_1 A_2 - A_2 A_1

Note that the commutator is itself a linear operator.

Kernel, range, and inverse

Kernel of an operator and injective maps

  1. Definition: Consider a vector space VV and a linear operator AA acting on it. The *kernelof of A$ is defined as

Ker(A)={vVAv=0}\text{Ker}(A) = \left\{ \ket{v}\in V | A\ket{v} = \ket{0} \right\}

As an example, consider V=C4V = \CC^4, and the operator

A=(1001100110011001)A = \begin{pmatrix} 1 & 0 & 0 & 1 \\ 1 & 0 & 0 & -1 \\ 1 & 0 & 0 & 1 \\ 1 & 0 & 0 & -1\end{pmatrix}

You can show that

Ker(A)={(0c1c20)   c1,2C}\text{Ker}(A) = \left\{ \begin{pmatrix} 0 \\ c_1 \\ c_2 \\ 0 \end{pmatrix}\ \ \forall\ c_{1,2} \in \CC \right\}

Note that this is a vector subspace of C4\CC^4. This is no accident, as we can prove:

  1. Theorem. For any vector space VV and linear operator AA acting on it, Ker(A)\text{Ker}(A) is a vector subspace of VV. (The proof in your problem set for the week).

  2. Definition. A linear operator AA acting on a vector space VV is injectove or one-to-one if

Av=Awv=wA\ket{v} = A\ket{w} \Rightarrow \ket{v} = \ket{w}
injective map

Finally, we note that these concepts can be extended naturally to maps between different vector spaces, though we will not use them in that context here.

  1. Theorem. A linear operator AA is injective if and only if Ker(A)=0\text{Ker}(A) = 0.

Range and surjective maps

  1. Definition. Consider a linear operator AA actig on a vector space VV. The range of AA can be fined as

Range(A)={Av vV}\text{Range}(A) = \{A\ket{v}\ \forall \ket{v} \in V\}

As an example consider V=C4V = \CC^4,

A=(0000111111110000) A = \begin{pmatrix} 0 & 0 & 0 & 0 \\ 1 & 1 & 1 & 1 \\ 1 & -1 & 1 & -1 \\ 0 & 0 & 0 & 0 \end{pmatrix}

Then

RangeA={(0c1c20)  c1,2C}\text{Range}{A} = \{ \begin{pmatrix} 0 \\ c_1 \\ c_2 \\ 0 \end{pmatrix}\ \forall\ c_{1,2} \in \CC \}
  1. Theorem: for any vector space VV and linear operator AA acting on that space, Range(A)\text{Range}(A) is a vector subspace of VV. I leave the proof as an exercise.

  2. Definition For a vector space VV, a linear operator AA acting on it is surjective or onto if Range(A)=V\text{Range}(A) = V.

  3. Definition For a linear operator AA acting on a vector space VV, the rangk of AA is Rk(A)=dim(Range(A))\text{Rk}(A) = \text{dim}(\text{Range}(A)).

  4. Rank-nullity theorem: For a linear operator AA acting on a vector space VV,

dim(Ker(A))+Rk(A)=dim(V)\text{dim}(\text{Ker}(A)) + \text{Rk}(A) = \text{dim}(V)

The proof can be found in Zweibach’s book Zwiebach, 2022, from which much of the material in this section is taken.

As an example, consider V=C4V = \CC^4,

A=(0100011001100100)A = \begin{pmatrix} 0 & 1 & 0 & 0 \\ 0 & -1 & 1 & 0 \\ 0 & 1 & 1 & 0 \\ 0 & -1 & 0 & 0 \end{pmatrix}

Then

Ker(A)={(c100c4)  c1,4C}\text{Ker}(A) = \{ \begin{pmatrix} c_1 \\ 0 \\ 0 \\ c_4 \end{pmatrix}\ \forall\ c_{1,4} \in \CC \}

This two-dimensional vector space has as a basis

v1=(1000) ;  v2=(0001)\ket{v_1} = \begin{pmatrix} 1 \\ 0 \\ 0 \\ 0 \end{pmatrix}\ ; \ \ \ket{v_2} = \begin{pmatrix} 0 \\ 0 \\ 0 \\ 1 \end{pmatrix}

Meanwhile

Range(A)={(c2c2+c3c2+c3c2)  c2,3C}\text{Range}(A) = \{ \begin{pmatrix} c_2 \\ - c_2 + c_3 \\ c_2 + c_3 \\ - c_2 \end{pmatrix}\ \forall\ c_{2,3} \in \CC \}

This space has as a basis

v3=(1111) ;  v4=(0110)\ket{v_3} = \begin{pmatrix} 1 \\ -1 \\ 1 \\ -1 \end{pmatrix}\ ; \ \ \ket{v_4} = \begin{pmatrix} 0 \\ 1 \\ 1 \\ 0 \end{pmatrix}

It is also two-dimensional, so the rank of AA is 2.

One can show that vk\ket{v_k} are all linearly independent. The rank of AA and the dimension of th ekernel are both 2, and 2+2=42 + 2 = 4 last time I checked, so the rank-nullity theorem holds.

Inverses

For all the below we will consider a veector space VV and a linear operator AA acting on VV.

  1. Definition. AL1A_L^{-1} is the left inverse if AL1A=1A_L^{-1} A = \bf{1}.

  2. Theorem: AL1A_L^{-1} exists if and only if AA is one-to-one.

Proof: First assume AL1A_L^{-1} iexists. Now consider v1,2\ket{v_{1,2}} such that

Av1=Av2A\ket{v_1} = A \ket{v_2}

Then

AL1(Av1Av2)=AL1=0=v1v2\begin{align} A_L^{-1}\left( A\ket{v_1} - A\ket{v_2}\right) & = A_L^{-1} \ket{\varnothing} = 0\\ & = \ket{v_1} - \ket{v_2} \end{align}

Thus v1=v2\ket{v_1} = \ket{v_2} and AA is one-to-one.

On the other hand, assume that AA is one-to-one. Now consider a basis vi\ket{v_i} of VV. Now any linear combination icivi=0\sum_i c_i\ket{v_i} = 0 if and only if ci=0  ic_i = 0\ \forall\ i. Since AA is a linear map, consider iciAvi=A(icivi)\sum_i c_i A \ket{v_i} = A (\sum_i c_i \ket{v_i}). Since AA is one-to-one, this can only be zero if icivi=0\sum_i c_i \ket{v_i} = 0 which as we have already said means that ci=0c_i = 0. Therefore, AviwiA\ket{v_i} \equiv \ket{w_i} is also a linearly independent set of vectors, and a basis for VV. We can then define

AL1wi=viA_L^{-1} \ket{w_i} = \ket{v_i}

Since wi\ket{w_i} is a basis, we simply demand that AL1A_L^{-1} is a linear operator, and this then determines the action of AL1A_L^{-1} on all vectors. Finally, we can see

AL1A(icivi)=AL1iciwi=iciviA_L^{-1} A \left(\sum_i c_i \ket{v_i}\right) = A_L^{-1} \sum_i c_i w_i = \sum_i c_i \ket{v_i}

for any cic_i, so AL1A=1A_L^{-1} A = \bf{1}.

  1. Comment. Assume AA has a left inverse. Then we can solve the equation

Av=wA\ket{v} = \ket{w}

for v\ket{v} given w\ket{w} by acting on both sides of the equation by AL1A_L^{-1} to ket v=AL1w\ket{v} = A_L^{-1} \ket{w}. However, we cannot necessarily check that this is a solution: we can act on both sides by AA but then Av=AAL1wA \ket{v} = A A_L^{-1} \ket{w} and we have no guarantee that AAL1=idA A_L^{-1} = \bf{id}. This becomes a particular issue for infinite-dimensional vector spaces. Consider the space of all polynomials

a0,a1,a2,k=0akxk\ket{a_0,a_1,a_2,\ldots} \Rightarrow \sum_{k=0}^{\infty} a_k x^k

Now if we define:

Sa0,a1,a2,=0,a0,a1,SL1a0,a1,a2,=a1,a2,a3,\begin{align} S\ket{a_0,a_1,a_2,\ldots} & = \ket{0,a_0,a_1,\ldots} \\ S_L^{-1}\ket{a_0,a_1,a_2,\ldots} & = \ket{a_1,a_2,a_3,\ldots} \end{align}

we can see quickly that

SL1Sa0,a1,a2,=a0,a1,a2,SSL1a0,a1,a2,=0,a1,a2,a3,\begin{align} S_L^{-1} S\ket{a_0,a_1,a_2,\ldots} & = \ket{a_0,a_1,a_2,\ldots}\\ S S_L^{-1} \ket{a_0,a_1,a_2,\ldots} & = \ket{0, a_1, a_2, a_3,\ldots} \end{align}

Thus SL1S=1S_L^{-1} S = \bf{1} but SSL1S S_L^{-1} does not act as the identity: it strips off the constant term from the polynomial.

  1. Definition. AR1A_R^{-1} is a right inverse if

AAR1=1A A_R^{-1} = \bf{1}
  1. Theorem: AR1A_R^{-1} exists if and only if AA is onto. We will not prove this here (the proof is complicated).

Again, we can furnish an example for which AAR1=1A A_R^{-1} = \bf{1} and AR1A1A_R^{-1} A \neq \bf{1}, by looking at the polynomials. Let

Ta0,a1,a2,=a1,2a2,3a3,TR1a0,a1,a2,=0,a0,12a1,13a2,\begin{align} T \ket{a_0, a_1, a_2,\ldots} & = \ket{a_1, 2 a_2, 3 a_3,\ldots}\\ T_R^{-1} \ket{a_0,a_1,a_2,\ldots} & = \ket{0, a_0, \frac{1}{2} a_1, \frac{1}{3} a_2,\ldots} \end{align}

We find

TTR1a0,a1,a2,=a0,a1,a2,TR1Ta0,a1,a2,=TR1a1,2a2,=0,a1,a2,\begin{align} T T_R^{-1} \ket{a_0,a_1,a_2,\ldots} & = \ket{a_0,a_1,a_2,\ldots}\\ T_R^{-1} T \ket{a_0,a_1,a_2,\ldots} & = T_R^{-1} \ket{a_1,2 a_2,\ldots}\\ & = \ket{0, a_1, a_2,\ldots} \end{align}

so TR1TT_R^{-1}T fails to be the identity operator because (once again) it strips off the constant term.

  1. Definition. A linear operator AA is invertible if AR1A_R^{-1}, AL1A_L^{-1} exists.

  2. Theorem. If AA is invertible, AL1=AR1A1A_L^{-1} = A_R^{-1} \equiv A^{-1}.

Proof. AL1AAR1=AL1=AR1A_L^{-1} A A_R^{-1} = A_L^{-1} = A_R^{-1} by associativity of operator multiplication.

  1. Theorem. AA is invertible if and only if it is one-to-one and onto.

  2. Theorem. Consider VV such that dim(V)=d<dim(V) = d < \infty. Then AA being invertible, AA being one-to-one, and AA being onto are all equivalent.

Proof. This follows from the rank-nullity theorem. We have shown that AA being invertible means it is one-to-one and onto. If AA is one-to-one, then dim(Ker(A))=0\text{dim}(\text{Ker}(A)) = 0, so by the rank-nullity theorem AA has rank nn; this means AA has all of VV as its image so it is onto. Similarly if AA is onto, the rank-nullity theorem shows that dim(Ker(A))=0\text{dim}(\text{Ker}(A)) = 0 so AA is one-to-one.

Matrix representations

I stated before that for V=CnV = \CC^n, all linear operators can be represented as matrices. In fact, this is true for any finite-dimensional vector space.

Consider a vector space VV with dimension nn. It will have basis I can label k\ket{k}, k=1,,nk = 1,\ldots,n. COnsider a linear operator AA. Since we can write any state in this basis, we can write

Ak=mAmkmA\ket{k} = \sum_m A_{mk} \ket{m}

AkmA_{km} here define n2n^2 complex numbers, and can be thought of as an n×nn\times n matrix.

Next, consider a general vector v=kvkk\ket{v} = \sum_k v_k \ket{k}. Then

Av=v=kvkk=kvkAk=k,vkAk\begin{align} A\ket{v} & = \ket{v'} = \sum_k v'_k \ket{k} \\ & = \sum_k v_k A \ket{k} \\ & = \sum_{k,\ell} v_k A_{\ell k} \ket{\ell} \end{align}

Since the expansion in a given basis is unique, we have v=Akvkv'_{\ell} = A_{\ell k} v_k.

In other words, given any basis, we can represent a vector as a set of nn complex numbers vkv_k. Of course we can arrange these numbers as a column vector

v(v1vn)\ket{v} \to \begin{pmatrix} v_1 \\ \vdots \\ v_n \end{pmatrix}

The action of the operator on that vector produces a new vector which can also be represented by a column vector

Av=v=(v1vn)=(A11A1nAn1Ann)(v1vn)A\ket{v} = \ket{v'} = \begin{pmatrix} v'_1 \\ \vdots \\ v'_n \end{pmatrix} = \begin{pmatrix} A_{11} & \ldots & A_{1n} \\ \vdots & \vdots & \vdots \\ A_{n1} & \ldots & A_{nn} \end{pmatrix} \begin{pmatrix} v_1 \\ \vdots \\ v_n \end{pmatrix}

In other words, every finite-dimensional vector space can be represented by Cn\CC^n (as the space of nn-dimensional column vectors), and every linear operator can be realized as an n×nn\times n matrix acting on column vectors.

You can show further (on your own time), that in the basis above, given two operators A,BA,B, (AB)k=AkmBm(AB)_{k\ell} = A_{km}B_{m\ell}; that is, operator multiplication is just matrix multiplication.

Changes of basis

The representation of vectors and operators as column vectors and matrices is basis-dependent. However, if we understand what the change of basis is, we can compute what a given vector and operator looks like in the new basis.

Consider two bases k,k~\ket{k}, \ket{\tilde{k}}. We can expand the basis vectors on the second in terms of the first:

k~=Uk\ket{\tilde{k}} = \sum_{\ell} U_{\ell k} \ket{\ell}

where UkU_{\ell k} are n2n^2 complex numbers. Note that UU is an invertible matrix as we can write

k=mVmkm~=mnUnmVmkk\ket{k} = \sum_m V_{mk}\ket{\tilde{m}} = \sum_{mn} U_{n m} V_{m k} \ket{k}

This only makes sense if UnmVmk=δnkU_{nm} V_{mk} = \delta_{nk}. We can run this in the opposite direction by swapping the tilde’dd and non-tilde’d bases in the above. The result is that VV is a left and right inverse for UU, and thus V=U1V = U^{-1} as a matrix.

How do operators look in the new basis? If we write

Ak~=A~k~A\ket{\tilde{k}} = \sum_{\ell} \tilde{A}_{\ell k} \ket{\tilde{\ell}}

then making the change of bases k~=Uk\ket{\tilde{k}} = U_{\ell k} \ket{\ell} on both sides, we find

A~k=Ukm1AmnUn\tilde{A}_{k\ell} = U^{-1}_{km} A_{mn} U_{n\ell}

Note here that the matrix UU maps between bases, and A,A~A,\tilde{A} are defined with respect to different basis.

The choice of basis is (at this stage) arbitrary, and one can ask what aspects of an operator are independent of the choice of basis. There are two particularly important quantities one can form:

  1. Given any basis k\ket{k} of VV, and an operator AA represented by the matrix AkA_{k\ell}, the trace of AA:

Tr(A)=kAkk\text{Tr}(A) = \sum_k A_{kk}

is independent of the choice of basis. To see this,

Tr(A~)=Umk1AklUlm=UlmUmk1Akl=δlkAkl=Akk=Tr(A)\begin{align} \text{Tr}(\tilde{A}) & = U^{-1}_{mk} A_{kl} U_{lm} \\ & = U_{lm} U^{-1}_{mk} A_{kl} = \delta_{lk} A_{kl} = A_{kk} = \text{Tr}(A) \end{align}
  1. Definition. The determinant of a matrix is also basis-independent. I will defer to your undergraduate linear algebra class to recall how to define the determinant of a matrix. The fact that it is basis independent comes from the fact that Det(AB)=Det(A)Det(B)=Det(BA)\text{Det}(AB) = \text{Det}(A)\text{Det}(B) = \text{Det}(BA). Thus

Det(A~)=Det(U1AU)=Det(UU1A)=Det(A)\text{Det}(\tilde{A}) = \text{Det}(U^{-1} A U) = \text{Det}(U U^{-1} A) = \text{Det}(A)

Eigenvalues and eigenvectors

  1. Definition. Consider a vector space VV and a linear operator AA acting on it. If there is a vector v\ket{v} and a complex number λ\lambda such that

Av=λvA \ket{v} = \lambda \ket{v}

then we call v\ket{v} an eigenvector of AA and λ\lambda the associated eigenvalue.

Note that for any eigenvector v\ket{v} and any constant cCc \in \CC, cvc\ket{v} is also an eigenvector with the same eigenvalue.

  1. Definition. Let vk\ket{v_k} be Kdim(V)K \leq \text{dim}(V) linearly independent eigenvectors of AA with the same eigenvalue λ\lambda. Then λ\lambda is called a degenerate eigenvalue, and vk\ket{v_k} span a degenerate subspace associated with A,λA,\lambda. We can show there is a maximal set of such linearly independent eigenvectors for a fixed eigenvalue. These fprm the basis of a subspace of VV valled the degenerate subspace associated with λ\lambda. The dimension of this subspace is called the geometric multiplicity of the eigenvalue.

  2. With a little thought you can show that the basis vectors for two degenerate subspaces associated to distinct eigenvalues are a linearly independent set of vectors.

  3. We can rewrite the eigenvalue equation above as

(Aλ1)v=0(A - \lambda \bf{1})\ket{v} = 0

This means that Ker(Aλ1)\text{Ker}(A - \lambda \bf{1}) is nontrivial, and therefore Aλ1A - \lambda \bf{1} is not an invertibel operator.

It is a basic fact that any operator AA is invertible if and only if Det(A)0\text{Det}(A) \neq 0. The reson is as follows. There is a standard formula in linear algebra for the inverse:

Amn1=(cofactor A)mnDet(A)A^{-1}_{mn} = \frac{(\text{cofactor}\ A)_{mn}}{\text{Det}(A)}

where the cofactor is the determinant of the (n1)×(n1)(n-1)\times(n-1)-dimensional matrix corresponding to AA with the mmth row and the nnth column deleted. This fails to be well defined precisely when the denominator vanishes.

Thus, to find the eigenvalues of a matrix, we can solve the equation Det(Aλ1)=0\text{Det}(A - \lambda \bf{1}) = 0. For VV with dimension nn, this is an nnth order polynomial equation. A simple example is a 2×22\times 2 matrix acting on C2\CC^2:

Det(Aλ1)=Det(aλbcdλ)=(aλ)(dλ)bc=λ2(a+d)λ+adbc\begin{align} \text{Det}(A - \lambda \bf{1}) & = \text{Det} \begin{pmatrix} a - \lambda & b \\ c & d - \lambda \end{pmatrix}\\ & = (a - \lambda)(d - \lambda) - bc = \lambda^2 - (a + d) \lambda + ad - bc \end{align}

This is equal to zero when

Det(Aλ1)=0λ=12(a+d)±12(a+d)24(adbc)=12(a+d)±12(ad)2+4bc\begin{align} \text{Det}(A - \lambda \bf{1}) = 0 \Rightarrow \lambda & = \half(a + d) \pm \half \sqrt{(a + d)^2 - 4(ad - bc)} \\ & = \half(a+d) \pm \half \sqrt{(a - d)^2 + 4 bc} \end{align}

Functions of operators

Lastly, we will need to define the concept of a function of an operator. Thet f(x)f(x) be a complex-valued function of a complex argument. How can we define f(A)f(A)? The clearest way is via a Taylor series. That is, if

f(x)=f0+f1x+f2x2+f(x) = f_0 + f_1 x + f_2 x^2 + \ldots

then we can define

f(A)=f01+f1A+f2A2+f(A) = f_0 \bf{1} + f_1 A + f_2 A^2 + \ldots

(We will find other ways to descibr operators once we have introduced Hermitian operators).

A particularly important example is the exponential of an operator:

exp(A)=n=01n!An\text{exp}(A) = \sum_{n = 0}^{\infty} \frac{1}{n!} A^n

It is straightforward to show that eλAeλA=1e^{\lambda A} e^{-\lambda A} = \bf{1}. First, we can use the power series representation to show that

ddλeλA=AeλA=eλAA ;\frac{d}{d\lambda} e^{\lambda A} = A e^{\lambda A} = e^{\lambda A} A\ ;

Thus

ddλeλAeλA=eλAAeλA+eλA(a)eλA=0\frac{d}{d\lambda} e^{\lambda A} e^{-\lambda A} = e^{\lambda A} A e^{-\lambda A} + e^{\lambda A} (-a) e^{-\lambda A} = 0

In other words, eλAeλAe^{\lambda A} e^{-\lambda A} is λ\lambda-independent, so we can find its value by setting λ=1\lambda = 1.

On the other hand, while for orderinary numbers a,ba,b, eaeb=ea+be^a e^b = e^{a + b} this fails to be true when re replac e a,ba,b with operators. To see this, we will loot at the first few terms in the Taylor series

eλAeλB=(1+λA+12λ2A2+)(1+λB+12λ2B2)=1+λ(A+B)+12λ2(A2+B2+2AB)\begin{align} e^{\lambda A} e^{\lambda B} & = (1 + \lambda A + \half \lambda^2 A^2 + \ldots)(1 + \lambda B + \half \lambda^2 B^2)\\ & = 1 + \lambda(A + B) + \half \lambda^2 (A^2 + B^2 + 2 AB) \end{align}

On the other hand,

eλ(A+B)=1+λ(A+B)+12λ2(A+B)2+=1+λ(A+B)+12λ2(A2+B2+AB+BA)+\begin{align*} e^{\lambda(A + B)} = 1 + \lambda (A + B) + \half \lambda^2 (A + B)^2 + \ldots \\ & = 1 + \lambda(A + B) + \half \lambda^2 (A^2 + B^2 + AB + BA) + \ldots \end{align*}

These two are only equal if AB=BAAB = BA, that is, if [A,B]=0[A,B] = 0.

In the intermediate case that [A,B]=c1[A,B] = c \bf{1}, with cCc \in \CC, one can (with a great deal of work) prove the very usefuil Baker-Campbell-Haussdorf formula:

eAeB=eA+B+12[A,B]e^A e^B = e^{A + B + \half [A,B]}
References
  1. Zwiebach, B. (2022). Mastering Quantum Mechanics: Essentials, Theory, and Applications. MIT Press. https://books.google.com/books?id=stKMEAAAQBAJ