In classical mechanics we saw that the states of the system were described as points on phase space. Observables of the system were functions on phase space, that could be described as generators of infinitesimal transformations.
In quantum mechanics, the space of ppossible states are described as vectors in a complex vector space (recall the discussion of photon polarization). Observables and generators of transformations are described as operators which map vectors to other vectors. More specifically, the correspond to linear operators which respect the linear structure of the vector space (addition and scalkar multiplication).
where a∗,b∗ are the complex conjugates of a,b respectively. This is important for defining adjoint vectors and normas of vectors.
Let V be a vector space over F. A linear operatorf is a linear map F:V→V.
For F=C, an antilinear operatorg is an antilinear map g:V→V.
Antilinear operators include the operator that implements time reversal, which is important in various condensed matter physics and particle physics contexts.
Abstractly we usually denote operators by capital letters A and the action of operators as A:∣V⟩→A∣v⟩.
More generally, scalar multiplication by any scalar in F is a linear operator.
For C=Cn, n×n matrices acting on column vectors by the usual rules of matrix multiplication are all linear operators. In fact one can represent any finite-dimensional vector space of dimension n by Cn, and every operator as a matrix acting on Cn.
A more interesting example: consider the (infinite-dimensional) space of all polynomials. We can represent these as
Relatedly, consider the vector space of square-integrable complex functions. You can convince yourself that multiplication by x and differentiation are linear operators. There are some subtleties here; it can be that these take the function out of the Hilbert space (eg multiplication can, if f falls off as 1/∣x∣ at infinity).
With a little work you can show that A1A2 is a linear operator. Note that if 1 is the identity operator, 1A=A1=A. Multiplication is associative: that is, if A3 is also a linear operator, ((A1A2)A3)=(A1(A2A3)). However, it does not have to be commutative. For example, for V=Cn with operators equal to n×n matrices, operator multiplication is just matrix multiplication which is certainly not commutative. Consider V=C2, and operators
Finally, we note that these concepts can be extended naturally to maps between different vector spaces, though we will not use them in that context here.
Theorem. A linear operator A is injective if and only if Ker(A)=0.
It is also two-dimensional, so the rank of A is 2.
One can show that ∣vk⟩ are all linearly independent. The rank of A and the dimension of th ekernel are both 2, and 2+2=4 last time I checked, so the rank-nullity theorem holds.
On the other hand, assume that A is one-to-one. Now consider a basis ∣vi⟩ of V. Now any linear combination ∑ici∣vi⟩=0 if and only if ci=0∀i. Since A is a linear map, consider ∑iciA∣vi⟩=A(∑ici∣vi⟩). Since A is one-to-one, this can only be zero if ∑ici∣vi⟩=0 which as we have already said means that ci=0. Therefore, A∣vi⟩≡∣wi⟩ is also a linearly independent set of vectors, and a basis for V. We can then define
Since ∣wi⟩ is a basis, we simply demand that AL−1 is a linear operator, and this then determines the action of AL−1 on all vectors. Finally, we can see
for ∣v⟩ given ∣w⟩ by acting on both sides of the equation by AL−1 to ket ∣v⟩=AL−1∣w⟩. However, we cannot necessarily check that this is a solution: we can act on both sides by A but then A∣v⟩=AAL−1∣w⟩ and we have no guarantee that AAL−1=id. This becomes a particular issue for infinite-dimensional vector spaces. Consider the space of all polynomials
so TR−1T fails to be the identity operator because (once again) it strips off the constant term.
Definition. A linear operator A is invertible if AR−1, AL−1 exists.
Theorem. If A is invertible, AL−1=AR−1≡A−1.
Proof. AL−1AAR−1=AL−1=AR−1 by associativity of operator multiplication.
Theorem. A is invertible if and only if it is one-to-one and onto.
Theorem. Consider V such that dim(V)=d<∞. Then A being invertible, A being one-to-one, and A being onto are all equivalent.
Proof. This follows from the rank-nullity theorem. We have shown that A being invertible means it is one-to-one and onto. If A is one-to-one, then dim(Ker(A))=0, so by the rank-nullity theorem A has rank n; this means A has all of V as its image so it is onto. Similarly if A is onto, the rank-nullity theorem shows that dim(Ker(A))=0 so A is one-to-one.
I stated before that for V=Cn, all linear operators can be represented as matrices. In fact, this is true for any finite-dimensional vector space.
Consider a vector space V with dimension n. It will have basis I can label ∣k⟩, k=1,…,n. COnsider a linear operator A. Since we can write any state in this basis, we can write
Since the expansion in a given basis is unique, we have vℓ′=Aℓkvk.
In other words, given any basis, we can represent a vector as a set of n complex numbers vk. Of course we can arrange these numbers as a column vector
In other words, every finite-dimensional vector space can be represented by Cn (as the space of n-dimensional column vectors), and every linear operator can be realized as an n×n matrix acting on column vectors.
You can show further (on your own time), that in the basis above, given two operators A,B, (AB)kℓ=AkmBmℓ; that is, operator multiplication is just matrix multiplication.
The representation of vectors and operators as column vectors and matrices is basis-dependent. However, if we understand what the change of basis is, we can compute what a given vector and operator looks like in the new basis.
Consider two bases ∣k⟩,∣k~⟩. We can expand the basis vectors on the second in terms of the first:
This only makes sense if UnmVmk=δnk. We can run this in the opposite direction by swapping the tilde’dd and non-tilde’d bases in the above. The result is that V is a left and right inverse for U, and thus V=U−1 as a matrix.
How do operators look in the new basis? If we write
Note here that the matrix U maps between bases, and A,A~ are defined with respect to different basis.
The choice of basis is (at this stage) arbitrary, and one can ask what aspects of an operator are independent of the choice of basis. There are two particularly important quantities one can form:
Given any basis ∣k⟩ of V, and an operator A represented by the matrix Akℓ, the trace of A:
Definition. The determinant of a matrix is also basis-independent. I will defer to your undergraduate linear algebra class to recall how to define the determinant of a matrix. The fact that it is basis independent comes from the fact that Det(AB)=Det(A)Det(B)=Det(BA). Thus
then we call ∣v⟩ an eigenvector of A and λ the associated eigenvalue.
Note that for any eigenvector ∣v⟩ and any constant c∈C, c∣v⟩ is also an eigenvector with the same eigenvalue.
Definition. Let ∣vk⟩ be K≤dim(V) linearly independent eigenvectors of A with the same eigenvalue λ. Then λ is called a degenerate eigenvalue, and ∣vk⟩ span a degenerate subspace associated with A,λ. We can show there is a maximal set of such linearly independent eigenvectors for a fixed eigenvalue. These fprm the basis of a subspace of V valled the degenerate subspace associated with λ. The dimension of this subspace is called the geometric multiplicity of the eigenvalue.
With a little thought you can show that the basis vectors for two degenerate subspaces associated to distinct eigenvalues are a linearly independent set of vectors.
This means that Ker(A−λ1) is nontrivial, and therefore A−λ1 is not an invertibel operator.
It is a basic fact that any operator A is invertible if and only ifDet(A)=0. The reson is as follows. There is a standard formula in linear algebra for the inverse:
where the cofactor is the determinant of the (n−1)×(n−1)-dimensional matrix corresponding to A with the mth row and the nth column deleted. This fails to be well defined precisely when the denominator vanishes.
Thus, to find the eigenvalues of a matrix, we can solve the equation Det(A−λ1)=0. For V with dimension n, this is an nth order polynomial equation. A simple example is a 2×2 matrix acting on C2:
Lastly, we will need to define the concept of a function of an operator. Thet f(x) be a complex-valued function of a complex argument. How can we define f(A)? The clearest way is via a Taylor series. That is, if
In other words, eλAe−λA is λ-independent, so we can find its value by setting λ=1.
On the other hand, while for orderinary numbers a,b, eaeb=ea+b this fails to be true when re replac e a,b with operators. To see this, we will loot at the first few terms in the Taylor series