From Picard iteration to Feynman path integral
Discrete path integral
As we all know, the Schrödinger equation is where is the state vector and is the Hamiltonian operator (may be time-dependent). This is an ordinary differential equation (ODE), so we can express its solution as a sum where each term in the sum is defined iteratively by (see also my past article) This iteration is called the Picard iteration, which is most known as a method to prove the Picard–Lindelöf theorem.
Let us actually write out the general term in this sum as where Now with the trick of time-ordering, we can rewrite this as where means to order the operators inside according to their time arguments. The factor of appears because there are ways to order time variables, but another way to see this is to note that the domain of integration is an -simplex, whose volume is of the corresponding -parallelotope. When we then sum over all , we get the time-ordered exponential The operator has a bunch of equivalent names, such as the time evolution operator, the propagator, the Green’s function, the Dyson operator, and the S-matrix (well, they are not entirely equivalent because they are used under different contexts).
The interesting part comes when we consider the matrix elements of and how they relate to the matrix elements of . We choose an orthonormal basis , and then insert a between each pair of Hamiltonian operators in the expression of : where we have abbreviated , and are a specific ordering of the integrated time variables. We can pull out the sum over intermediate basis states and call the summand a contribution from a walk 1 from to . In other words, where the sum is over all walks of length from to , and Now, the matrix elements of the full propagator is then the sum over contributions from all walks: We can imagine a “Hamiltonian graph” formed by taking the basis states as vertices and the Hamiltonian matrix elements as edge weights (the weight of the directed edge from to is ). Note that an edge can be a self-loop. Then, the propagator contribution from a walk is given by integrating the product of all the edge weights along the walk (for a trivial walk, which has zero length, the contribution is simply ). This formulation may be called the discrete path integral. There is a paper on arXiv that delivers the idea of the Hamiltonian graph. Its difference from the current article is that it only focuses on time-independent Hamiltonians and that it treats self-loops separately instead of just like normal edges. The following two sections (excluding self-loops and the Feynman path integral) largely follow from the ideas from this paper.
Excluding self-loops
In some cases, it may be hard to consider self-loops on the Hamiltonian graph. We may then benefit from counting only walks without self-loops. However, the contributions from each walk will now be different because we have to account for the same walk with self-loops inserted at various positions. In other words, for a walk without self-loops, instead of contributing , we now want to find the contribution that sums over all ways to insert self-loops into : where is the number of self-loops inserted at the vertex , and is an abbreviation of this walk:
I will show that we can find an expression for for the case when the Hamiltonians at different times commute with each other. In this case, we have Therefore, by the definition of , we have For abbreviation, for the rest of this section, we denote .
Now, we use a trick to replace the factorial in the denominator with an contour integral: where the contour is a counterclockwise simple closed curve around the origin in the complex plane. We then have We have thus separated the sums over different . Each sum is a geometric series that converges when we choose the contour large enough, so where the last step used the residue theorem for each pole at . This expression is exactly the expanded form of the divided difference of , often denoted . Therefore,
The discrete path integral can then be rewritten in a form that only involves collecting contributions from walks without self-loops:
Feynman path integral
This discrete path integral formulation of the propagator already looks similar to the Feynman path integral, but we have to go a step further to take the continuum limit to actually get there. For simplicity, I will only consider a particle with unvarying mass moving in a time-independent potential in one dimension. Its Hamiltonian is , and the orthonormal basis is chosen to be the position basis, also denoted as .
The more standard way to derive the Feynman path integral is to slice the time integral in Equation 1, to express the total exponentiation as a product as many small exponentiations, and then to insert between each pair of exponentiations (see, e.g., chapter 6 of Quantum Field Theory by Mark Srednicki). However, this approach does not make its connection to the discrete path integral clear. Instead, we will discretize the position space into a lattice with spacing , use the discrete path integral formulation on this lattice, and then take the continuum limit at the end. Now, instead of , we have . Each basis vector now has two nearest neighbors and .
In the position basis, the kinetic part of the Hamiltonian is a second derivative operator. From numerical differentiation, we can approximate it on the lattice as Therefore, the discretized Hamiltonian is Its matrix elements are then Conceptually, it consists of on-site energy and nearest-neighbor hops. The on-site energy looks bothersome, but we can remove it if we only consider walks without self-loops. Equation 2 becomes where is the imaginary time, and is defined for the same abbreviation reason as the previous section. The terms proportional to in the multiplicant are omitted because we only consider walks without self-loops. The rest of this section is done under a Wick rotation so that is assumed to be a positive real parameter.
First, let us tackle the divided difference . Define , where is the mean of . Then, is of order unity (while is of order , which is much larger than unity for small ). Then, from the expanded form of the divided difference, we can easily get Recalling how we initially derived the divided difference, we have When is large (the reason of which will be explained in a minute), we have etc., while is of the order of unity, so we only need to consider the terms with the lowest . The leading term is the term with , which is trivially , so we have This contributes to the potential part of the action.
Substitute Equation 4 into Equation 3, and we have where is a large positive number when is small. Observe that the factor is the probability mass function of the Poisson distribution with mean evaluated at . When is very large, the Poisson distribution can be approximated by a delta distribution because the standard deviation is much smaller than . In other words, . Therefore,
Imaginary parameter Poisson distribution
It was this point that got me thinking the most when I originally tried to derive the Feynman path integral without the Wick rotation. While the approximation is valid, the problem is whether we can likewise say (or ). While it is true that the left-hand side has a very large magnitude when so that it dominates the sum, it does not actually approximate the right-hand side because the right-hand side is of order unity and is real. In fact, the summand is rapidly oscillating when is near , so the numbers of different phases actually cancel each other out and give a number with small magnitude in the end.
If you actually try to walk through the calculation without the Wick rotation, you will find that what you need to justify in the end is something like this (there are some other factors dependent on in the summand, but we can remove them by some techniques, so let us ignore them for simplicity): where and are both large positive numbers. This is unfortunately false, neither in magnitude nor in phase, and not even up to an overall factor.
While it is true that a lot of things can be carried over by analytic continuation, which is the reason why the Wick rotation can give the correct result in many cases, you can do the analytic continuation only if every step you take is actually analytic. Having an approximation based on the magnitude of each summand is not analytic because a fast oscillation can change the result drastically. Therefore, I am not satisfied with this derivation with the Wick rotation, but I have not found a better way to do it yet.
If the factor involving were not there in the summand, the sum of products is exactly the probability that a random walk starting at ends at after steps, where at each step the walk moves to the left or right nearest neighbor with equal probability . Instead of considering one random walk with steps, we can consider random walks with steps each, where both and are large. Because is large, we can approximate the distribution of the position at the end of the th random walk as a normally distributed random variable with variance . Note that even though is large, is still very small, so the majority of contribution in the sum only comes from those paths where does not differ too much from the of its corresponding part of random walk. Therefore, if we fix a set of , the factor involving in the summand can be approximated by replacing with for all in the th random walk segment. We can then pull this factor out of the sum over (but still inside the integral over ). Therefore, we get where the extra factor of in the multiplicant comes from converting a probability density to a probability (since the probability that the position ends up at the lattice site is times the probability density at ).
Combining the product of exponentiations into the exponentiation of a sum, we have If we introduce the time step , for large , we have Similarly, for the potential part we introduce the time step 2, Here, and are the position and its time at time for a particle undergoing these random walks. Therefore, we have
Finally, simply revert the Wick rotation by substituting and rewrite the integral over as a path integral over all paths . Then, we get the Feynman path integral expression of the propagator: This completes the derivation.
You may wonder why I use the word “walk” while the resultant thing is called a “path” integral. This is just because “walk” is the correct term in graph theory that describes the object we use here, and the sum over all the walks is called a “path” integral because it is what physicists call. In graph theory, however, a path is a walk in which all vertices are distinct.↩︎
Do not ask my why it is . It is not important.↩︎
