The next “non-perturbative” method we will discuss is the Wentzel-Kramers-Brillouin-Jeffreys approximation (sometimes the “J” is not present). This is a technique for solving wave equations, focusing on the case that the relevant energies are “large” in a sense that we will discuss.
This method is quite powerful and has uses well beyond quantum mechanics. It is the underlying scheme by which we justify ray-tracing methods for solving differential equations, such as geometric optics. For a good general discussion, I recommend Bender & Orszag, 1999. (A wonderful book for many subjects).
In the case of quantum mechanics, the other beautiful outcome of the WKB approximation is that we can begin to see the emergence of classical mechanics. We will discuss this fact once we have gone over the basics of the approximation.
While this is not a perturbative method, as it applies even when there is no “nearby” exactly solvable Hamiltonian, it is an asymptotic method. That is, there is a small parameter, and a systematic way of expanding the solution in that small parameter. The lowest-order solution is singular: in particular it is non-analytic in the small parameter. But there is an expression for it that can often be solved even when the underlying Schroedinger equation cannot be.
The short-wavelength limit¶
Generalities¶
We are going to focus here on the one-dimensional time-independent Schroedinger equation:
Crudely speaking, we are interested in highly excited states, in a way we will try to make precise. Consider in particular a bound state problem as illustrated below.
Inside the potential well, at high eneough energy there will be many nodes and the wavfunction will oscillate rapidly. In this region, the approximation works when the number of nodes is . If the potential is not varying too rapidly, it will be approximately constant over the scale of separation of the nodes. Locally near these nodes, the solution will then look like a plane wave (solution in a constant potential) with wavenumber
Similarly, if the energy is far enough below the potential, then the exponential falloff will be rapid compared to the rate of change of the potential and cam be characterized locally by a decay rate
which we can think of as an imaginary wavenumber.
If this approximation is close to true, we can approximate the wavefunction well inside the potential well with
where is the classical momentum at fixed energy. Here is a choice, that can be absorbed into . Well below the barrier, the solution is:
where is a sort of imaginary momentum.
The approximation can be expected to break down when the energy is roughly equal to the potential energy, so that the variation of the wavefunciton is not rapid. This is the region around the classical turning points. However, in this region, we can approximate the potential as a linear one, and the Schroedinger equation with a linear potential does have a solution in terms of Airy functions. If the WKB approximation starts to be good while the linear approximation is still valid, then we can use this fact to match the Airy functions to the WKB solutions and get a complete solution, including an approximation to the energy eigenvalues.
To get an estimate of when this approximation words, we can ask that the change in the wavelength over a distance of order the wavelength is small compared to the wavelength. This becomes
So the change in the potential over a wavelength is small compared to the kinetix energy. We will see how this conditions emerges from a more systematic treatment.
The semiclassical expansion¶
Let us begin with the ansatz
So far this is just a rewriting, not an approximation. I am using the letter to foreshadow the relationship of the WKB approximation to the emergence of classical mechanics.
Putting this into the TISE, we get
This is still an exact equation. But the approximation (4) arises if we are allower to ignore the final term. Since this aproximation seems tied up in classical mechanics, and the last term is proportional to , one often says that we will perform a “small-” approximation. This is not quite correct. What we will be demanding is essentially that .
With this in mind, we can still write a formal expansion
and solve the Schroedinger equation order by order in . We will still need to check that . Note that in terms of the wavefunction this is a singular expansion; the lowest-order piece is nonanalytic in .
Plugging (9) into (8) and keeping only the -independent term, we get
Those who have had a good classical mechanics course will recognize this as the Hamilton-Jacobi equation, where is the classical action. We will return to this interpretation below, which is central to ray tracing/geometric optics/the method of characteristics. The solution is
where again, is an integration constant. We have two solutions to and therefore for . At this level of approximation we can consider linear combinations of these two solutions. Note that what we have done is true formally whether is positive or negative. In the former case, we recover (4). In the latter case, becomes imaginary, and we recover (5).
Our next step is to find . The terms in (8) are:
Thus, keeping only terms that are nonvanishing as , we have
where . This prefactor has a nice semiclassical interpretation, especially for bound states. We know is the probability density for the particle at position . If the particle is in a potential well it will oscillate with period . The fraction of time the particle will spend in a region is the time spent in that region, which is
Finally, we need to check on our approximation. Thius should be done at every order but we will here simply ask that in (8) the last term is small compared to the rest: that is, that
This means
which is essentially the condition we already derived from a more hand-waving argument. Here we introduce for notational ease in the example below.
Just to get a sense of the nontriviality of these conditions, we can look at a 1d case with the potential with . If we fix the total energy , we find
For fixed this is always valid as as the particle becomes the free particle for which WKB is exact. As , when the potential energy dominates is large compared to , we have:
Thus WKB works well when . For WKB breaks down near the origin. (Note that for the energy eigenstates are unbounded below, with being an interesting marginal case, as we have seen in a prior problem set).
You might fairly ask why we chose the scaling we did. Why is the leading singular behavior and not or or some such? Why do we expand in powers of rather than , ? This is determined by the power of that sits in front of the kinetic term. If we demand that all of the terms in the Schroedinger equation appear at the same order, including the energy, you can convince yourself tha this is the right scaling. In general you have to deduce the correct scaling; and a given problem may support various scalings for different energies and momenta. The book Bender & Orszag, 1999 has a good discussion of these issues. This paper of which I was a co-author encounters this issue in a situation where the resolution is nontrivial.
Bound state problems¶
We now consider particles in some potential well, such that . In the figure below, we have broken up the axis into three regions. In regions I, II, III the WKB approximation is expected to hold. For large (in regions I and III), we assume the energy is far enough below the barrier that the length scale giverning the exponential falloff of the wavefunction is small compared to the length scale over which the potential varies. In region II we assume that the energy is large enough that the wavefunction rapidly oscillates, with a space between nodes so small that the potential does not vary appreciably in this range.
The WKB approximation is expected to break down at the classical turning points > Here , , and we can see that the condition (6) breaks down. Generally, however, the potential can be well-approximated by a linear potential; the resulting Schroedinger equation is exactly solvable. If we are fortunate, the regime in which the linear approximation holds will overlap with the regime in which the WKB approximation holds, as illustrated below. In this case we can match our WKB functions to the solutions near the turning points.
Let us first turn to the solutions in the WKB regions. In region I, the solution is
Here is arbitrary. This solution will be exponentially decaying as : we have set to zero an exponentially growing component.
In region II the WKB solution looks like
While in Region III the WKB solution looks like
which is exponentially decaying.
As with the other bound state problems we have studied, we have 4 complex unknowns, with one complex constraint due to the combination of normalization and choice of overeall phase. We will find four constraints due to matching the WKB solutions to the equations near the turning points.
Let us study the turning point . Expanding in a Taylor series,
We assume that teh higher order terms are small in a region around that overlaps with the region of validity of the WKB approximation. In general, of course, this needs to be checked.
Now we examine the Schrodinger equation. By definition, , so these terms drop out. Furthermore, we define , with to be determined. Nondimensionalizing our equations are almost always a good idea -- we will discover important characteristic scales, and we will reduce the equations to ones that we can look up or plug most easily into a computer. Further multiplyng the equation by , the resulting equation is:
We set , to find
This differential equation has known solutions in terms of Airy functions, whose properties you can look up in many places such as the Digital Library of Mathematical Functions (see Chapter 9). There are two independent solutions , and a general solution has the form
At large , the WKB approximation works well for these functions (you can and should deduce this from the conditions we have discussed). Their asymptotic properties are known:
You can check easily that these result from applying our WKB solutions to regions to the linear potential.
Normalizability then demands that we choose solution about the turning point . About we can map the problem to the one we solved by letting , so that the proper solution is .
We then need to match the solutions in region II. Performing the expansion of near the turning point,
where . The minus sign appears because we get inside the square root; the upshot is an overall minus sign when we do the integral. We set to find that our solution in region II matches if we adjust such that
Note the integral is from to ; this takes care of the minus sign we introduced.
On the other hand, the same matching condition applied to the turning point gives us
These two need to be equal, which means the cosine terms need to be equal. Since the cosine terms are odd, we can write
The arguments now need to be equal up to a factor of (the overall sign difference if is odd can be absorbed into the overall relative sign of ), so that
which becomes
This is the Bohr-Sommerfeld quantization condition. It was generalized to higher-dimensional situations in various works by Einstein, Brillouin, and Keller (only applicable to energies for which classical trajectories live on tori in phase space; a story that will take us too far afield, however).
Example 1: the Simple Harmonic Oscillator¶
In this case the turning points are at . The quantization condition is
Now we can substitute and set ; we find
But this is the exact answer! This is a general feature of Hamiltonians quadratic in their arguments -- the WKB approximation gives exact expressions for the eigenvalues. It also, through the path integral, gives an exact expression for teh propagator, as we will discuss. It does not, however, give the right expressions for the energy eigenstates. This is a longer story; the chapters in Shankar on path integrals and their relation to WKB contains the required pieces of this puzzle.
Example 2: the Linear potential¶
Consider instead . If we go through the same exercise, we find
The wavefunctions, as it happens, are amenable to an exact solution using Airy functions, and the resulting eigenvalues can be computed. For the ground state , the WKB expression is too high by . for the first excited state, the result is too high by . For the second excited state, the result is low by . (It is worth thinking about why the WKB answers will always be larger than the exact answer, but at higher there is no such restriction).
The Hamilton-Jacobi equation and geometric optics¶
We found that at lowest order in the WKBJ approximation,
Here we have written down the equation for an -dimensional nonrelativistic potential problem; the derivation is essentially the same.
As it happens this is a well-known equation in classical mechanics, the Hamilton-Jacobi equation. Since this course teaches classical as well as quantum mechanics, I will break here and discuss the meaning of this equation in classical mechanics.
Hamilton-Jacobi from least action¶
Let us return to the principle of least action and consider it in phase space form. Starting with , we recall that to first order in variations of ,
In the usual Euler-Lagrange equations we fix the initial and final positions. Here we want to understand evaluated on classical paths as a function of , given some fixed initial condition . Since we are evaluating along classical paths, the integral vanishes, the integrand just being the classical erquations of motion, and we are left with , or
where is the conjugate momentum. Here since is fixed and given, is completely determined.
Now by definition, by definition. On the other hand, using the chain rule,
Since , we have
Thus if we consider as a function of , its variation under infinitesimal variations of its arguments is
We have also shown that
when evaluated along classical trajectories. This is the Hamilton-Jacobi equation; the function is called Hamilton’s principle function.
We have shown that by knowing the classical trajectories given fixed initial conditions, the equations (38) and (40) furnish a solution to the Hamilton-Jacobi equations. Furthermore, this solution is the classical action as a function of the endpoint given . Returning to the quantum problem, if we consider the time-dependent Schroedinger equation
and write \psi = e^{i\frac{S(q,t)}{\hbar} with , then the leading order equation is the full Hamilton-Jacobi equation.
The equation (36) is consistent with the full Hamilton-Jacobi equations if we set
where is called Hamilton’s characteristic function; here , or . The solution is also consistent with the classical ewuations of motion for the family of trajectories with fixed energy. The functional form will be different from the solution with fixed initial conditions, as the corresponding trajectories will have varying energy (consider the 1d case of force-free motion where we fix and vary given an initial condition; this will be a family of trajectories with increasing velocity and thus energy).
What we are seeing is that different solutions to the H-J equations with different specified constants yield different families of trajectories. To understand this we consider a different perspective on these equations.
Canonical transformations in phase space¶
Last semester we briefly touched on canonical transformations as transformations in phase space which preserve the Poisson brackets and the Hamiltonian structure of the problem. That is, start with some system for which the phase space variables are , , such that , and . Assume further there is a Hamiltonian such that the system satisfies Hamilton’s equations. A canonical transformation is coordinate transformation in phase space to new coordinates , such that
and such that there is function for which Hamilton’s equations for become, under the coordinate transformation, Hamilton’s equations for :
Following the previous discussion, these are consistent with the trajectory being the minimum of the phase space action
or in differential form, .
The variation of yields the same equations of motion if where is a function of coordinates, momenta, and time; integrating this along a trajectory, it will only contribute to the endpoints and thus not appear in the equations of motion. (Note that in principle stationarity works if for some constant . But we can scale away, by rescaling and rewriting the form of , so we will ignore this. (See chapter 10 of Goldstein et al., 2002 for a discussion of this point).
As it happens, however, determines the coordinate transformation. That is, if we write
or equivalently
The differential version treats as the independent coordinates. Thus, we have
If we specify , then, the equations above give , , . We can then solve for in terms of .
On the other hand, writing , we can instead consider the equation
If we set we get a different set of coordinate transformations determined by , with
We can consider similar generating functions that are functions of or of . These are typically classified into four types, dependening on which pair of old and new variables they are explicitly functions of.
Let us therefore consider coordinates that are constant in time along trajectories. These could be the initial conditions or any other set of bits of data that label the trajectories. The corresponding Hamiltonian is therefore constant by Hamilton’s equations and can be set to zero. Then we are left with
where we have set to tie this to the discussion in Section 9.1 of Goldstein et al., 2002. The variables appear as explicit constants of motion; we are guaranteed that is also constant. The equation above becomes the Hamilton-Jacobi equation
In the original derivation we specified . But as long as we have a complete solution with independent integration constants, then we can invert the coordinate transformations to find , that is, we generate a trajectory.
Canonical transformations and the HJ equation¶
In this context, if the Hamiltonian is time-independent, we can choose a solution of the form . now becomes one of the constants of motion: the resulting equation
A complete solution will require additional constants of motion.
The simplest interesting example is the 1d simple harmonic oscillator. Considering solutions of fixed energy, we have
We can do an integral to find
up to a constant of integration which we can absorb in to without changing the fact that solves the H-J equation.
By the above discussion, is a constant of motion, which we will call , so that
is constant; again the constant of integraion unspecified by the indefinite integral can be absorbed into . The result is
or
This is the general solution; the constants of motion are and the time at which . The latter can be shifted away with a shift in time.
The method of characteristics¶
The flip side of this, relevant for the quantum problem, is that if we have a family of solutions to the equations of motion, we have a solution to the Hamilton-Jacobi equation. That is, let us assume we know (the initial condition) at some time . This yields . At each initial point we can thus generate a trajectory . Along each trajectory, the partial derivative becomes a total derivative, so we have . is a known function and the equations of motion yield known functions , so we have which can be integrated, perhaps on a computer. We then know along the family of trajectories we have computed. This is a specific example of the method of characteristics. The process of turning the wave equation into an equation of Hamilton-Jacobi type, and then solving the katter via the method of characteristics, is a specific example of ray tracing. A similar procedure for exlectromagnetic waves, for example, moving through a medium with variable index of refraction, yields the geometric optics approximation, where the individual ``rays" are the characteristics of an associated Hamilton-Jacobi-like approximation to the full wave equation.
A simple example is the case of the free particle, for which
The characteristic equations/Hamilton’s equations are , {\dot p} = 0x(t) = x_0 + p_0 t/mt = 0S = \half \gamma x_0 ^2p_0 = \gamma x_0x(t) = (1 + \gamma t/m) x_0$.
Now along this trajectory,
Now along the trajectory , , so or
You can check that this solves the Hamilton-Jacobi equation.
WKBJ as a semiclassical limit¶
The fact that the classical equations omf motion arise from the WKBJ approximation motivates the term semiclassical limit. You might think this is because it is an limit; again, that is not quite right, since (as I keep ranting on and on about) is a dimensionful parameter.
I will sketch out a couple of ways to think about this limit. These are only sketches. Fleshing out the arguments in detail is non-trivial!
Comment: large quantum numbers¶
In general the story is that . We can see what this means in terms of the simple harmonic ocillator. If , wheer is the excitation level of a state, we can guess that . So is the limit; that is, we have a limit of large quantum numbers.
Anther way we might see classical mechanics emerge is through Ehrenfest’s theorem. For the potential problem, this states
This is exactly the classical equations of motion for if
As it happens this is trivially true for the SHO, for which the WKB approximation is known to be exact. More generally, if we write , we need that . The latter measures the spread of the wavefunction; we need this to be small compared to the range of variation of the position variable. This should be related to the kind of high-energy limit we are discussing in which the wavelengths used are small compared to the range of variation of the potential. We can realize this by constructing coherent states which realize these conditions explicitly. In view of time, however, I will press on.
Comment: the path integral picture¶
Let us return to the Feynman path integral for solutions to the time-dependent Schroedinger equation:
This would yield the WKB approximation if the dominant contribution to the integral was . As it happens, this emerges fairly naturally. I will sketch out a hand-waving description here.
Let us first consider the general 1d integral
As , if varies nontrivially with , eg , the integrand will oscillate rapidly and the integral will vanish. The leading contributions to this integral occur when changes slowly, that is, near . In this region, we can write . If we insert this expansion into the integrand,
(we expect the cubic term in the integrand to drop out as it is odd and the exponential even). The resulting leading-order approximation is known as the stationary phase approximation (and is related to the saddle-point approximation for more general integrals in the complex plane).
If we assume that this approximation applies to the Feynman path integral, we should replace the point by the classical trajectory about which is stationary. We are thus led to
times a prefactor. Here is the classical action, and we have shown that it solves the Hamilton-Jacobi equation.
In general, this works when . In explicit examples, you can show that this ratio controls the corrections to the semiclassical approximation, whether you work in the path integral formulation or with the Schroedinger equation. Such large actions arise from highly excited states.
- Bender, C. M., & Orszag, S. A. (1999). Advanced Mathematical Methods for Scientists and Engineers I: Asymptotic Methods and Perturbation Theory. Springer. https://books.google.com/books?id=-yQXwhE6iWMC
- Goldstein, H., Poole, C. P., & Safko, J. L. (2002). Classical Mechanics. Addison Wesley.