3 Interacting scalar field
Now that we understand free quantum fields, we can turn our attention to interacting fields. Unfortunately, there are only a few QFTs that permit an exact analytic solution such as the Schwinger model, a two-dimensional description of a photon and fermion. Most of these are only useful in very limited circumstances or as toy models. Another approach is to solve the theory numerically, usually by placing it on a finite lattice. While this allows the study of realistic models like quantum chromodynamics (QCD), it is extremely complicated to do from scratch.
Here we will instead focus on perturbation theory, i.e. we will consider the interaction to be a small perturbation on top of the free field which we already understand. This means we split the full Lagrangian
| (124) |
into a free theory and an interaction term which is hopefully small. Perturbation theory allows us to study theories like quantum electrodynamics (QED) or QCD (at least in the high-energy limit). The calculation of the anomalous magnetic moment of the electron I have mentioned at the beginning of the course is (almost completely) done this way. Similarly, calculations that are used at Large Hadron Collider (LHC) and that, for example, helped discover the Higgs boson in 2012, are also mostly done using perturbation theory. This is valid because the coupling strength between different particles, i.e. , is small in these regimes. In the example for , we expand in and the theory value includes effects up to . At the LHC we usually expand in the strong coupling and a few cutting-edge calculations have reached accuracy.
In this chapter we will only consider the self-interaction of a scalar particle . Nature has a fundamental scalar particle, namely the Higgs boson, and the theory we develop here can (almost) be used to describe the Higgs boson. The difference is that the real Higgs field is not a single real-valued field (like the one we discussed) but a doublet of two complex-valued fields. The relevant part of the Lagrangian is usually written as (like on that mug)
| (125) |
We will modify this slightly for our single real-valued scalar field
| (126) |
I want to stress that you could repeat everything we are about to do for the more general case of the Higgs boson and I would encourage you to try. This theory is often called the model due to its interaction term. Another related model would be the theory which would have which can serve as a simplified model of QED.
It is still possible to derive CCRs for an interacting Lagrangian but we will not be able to write in terms of ladder operators because this relied on the free field’s EoM. The important concept behind perturbative QFT is that the interaction is not just small but usually also short-ranged. This means that at the beginning and end of the experiment, we can consider the particles practically free. For example at the big LHC which has big detectors, the actual perturbative interaction takes place in a region that is smaller than (based on a typical hard interaction scale of ). For the particles involved in the collision are just too far away from each other to feel each other’s influence, meaning that we can use the free-particle solution in this regime. We now need to try and find a way to formalise this.
3.1 Interaction picture
In Section 2.1 we have seen the Heisenberg and Schrödinger pictures. In the former, the time dependence fully resides in the operators while in the latter it is in the states. We will now develop a third picture, the interaction picture that splits the time dependence between the two: the operators will follow the free Hamiltonian ’s Heisenberg equation (70) while the states have a time dependence from the interaction Hamiltonian . From (124), it follows immediately that
| (127) |
Recall how we wrote
with
| (66) |
Rather than taking the full in (66) we now only use the free to define
| (128) | ||||
| (129) |
The time evolution of is still governed by the free Heisenberg equation (70)
| (130) |
while the states follow a modified Schrödinger equation
| (131) | ||||
where we have used the Schrödinger picture Schrödinger equation for and defined the interaction picture interaction Hamiltonian .
These rules tell us how to transform from the Schrödinger picture to the interaction picture. We still need to translate between the Heisenberg and interaction pictures. At some instant , we define the pictures to be identical. At any other time , we translate via the Schrödinger picture
| (132) |
It is easy to see that the time evolution of is
| (133) |
3.2 The matrix
We are now ready to describe a scattering process. The time evolution of the field operator is given by . Let us take the limit , well before the scattering takes place. This is where we require the interaction picture and Heisenberg pictures to agree. The field operators in this limit are just the free field which we refer to as the “in” contribution and write in terms of ladder operators
| (134) |
Note that still has the time dependence from the free Hamiltonian. This is the case despite the limit since there is still the dynamics of the free field. At a later time, including all the way through the scattering, we have
| (135) |
Experimentally, we do not observe but rather the outcome of the scattering in the far future
| (136) |
This “out” field is once again free and we can write it again in terms of ladder operators
| (137) |
Both and are free field solutions but they are different free field solutions. This is because, due to the scattering, the ladder operators and are different. We refer to these states as the asymptotic states to indicate the limit .
The relation between the two sets of asymptotic states is governed by the time evolution operator . For simplicity, let us define the matrix
| (138) |
Specifically,
| (139) |
This means that also transform between in and out states. We begin our experiment with a prepared in state by applying on the vacuum. During the scattering this gets transformed into an out state which is made up through . We will use a subscripts and to indicate the ladder operators we have used. These two states are related through the matrix
| (140) |
To understand the scattering process we first need write in terms of the basis of the out states
| (141) |
Experimentally we will measure the transition probability between our prepared and a given out state
| (142) |
This means our goal will be to find an expression for the matrix.
To do this, we would first need to find to use (138). The definition (132) is not very helpful because of how complicated it is. Instead, we will use the differential equation (133) which defines this solution in the first place. Integrating from to
| (143) |
Note that, because we fixed , the interaction Hamiltonian is defined in terms of in states. Unfortunately, this expression still has a on the right-hand side so let us iterate this
| (144) | ||||
| (145) |
Per our construction of the interaction picture, only contains the interaction and not the dynamics of the free field. For example, in the theory we defined in (126), we had . Since we further requested that the coupling of interaction is small, we would be justified to assume that is smaller than . We therefore often choose to truncate the summation at a finite (in practice is this rarely more than or because of the complexity of the calculation).
One important feature of the iterated integrals in (145) is that we go further into the past in the product of since
| (146) |
which makes for awkward integration boundaries. Instead, let us define the time-ordered product similar to the normal ordering we have used before. Specifically,
| (147) |
where is the permutation of such that time is ordered, i.e.
| (148) |
We can now change the integration domain to for each integral at the cost of a factor of . Explicitly for , we split the integral into two equal pieces and swap
| (149) |
Doing the same for all orders, we can rewrite as
| (150) | ||||
| (151) |
Here we have introduce the time-ordered exponential as a short-hand. We can make one more simplification by taking the limit and by realising that
| (152) |
to arrive at our most compact solution for the matrix
| (153) |
with the interaction part of the action . Looking at this you might think there is some deep interpretation of this expression in terms of the action, similar to the principle of least action we had in the classical case. And you would be right to think this, it is possible to perform the entire quantisation procedure for free and interacting fields by defining the path integral
| (154) |
which integrates over all possible field configurations and weights them according to . In the classical limit only the field configuration of the least action contributes to the integral. While very elegant, the path-integral formalism is more complicated since we would have to define what means. Therefore, we will not use this method going forward instead relying on the canonical quantisation we have used so far.
3.3 Wick theorem
If we want to make a prediction about a scattering process we need to calculate matrix elements to a given order. This means calculating correlators like
| (155) |
Since both and the external states involves a number of fields, we want to be able to calculate general objects like
| (156) |
We have already seen a simple case of this with the Feynman propagator in (123). To make calculating this easier, let us define
| (157) |
This decomposition into positive () and negative () frequency modes is helpful because
| (158) |
It also makes it easier to define normal ordering which moves the , and therefore , to the right of the , and therefore . It follows that the vev of a normal-ordered list of fields is zero
| (159) |
To see why this is so useful, consider again the two-particle case that we considered when defining the Feynman propagator. For ,
| (160) |
Every term expect for the commutator is now a normal-ordered product of interaction-picture operators. The commutator is the only term with a non-vanishing vev. Because the interaction-picture fields follow the time evolution of the free Hamiltonian, we can use what we discovered in the previous section. Especially, we can use that the commutator corresponds to the Feynman propagator .
To simplify our notation, let us define a Wick contraction as
| (161) |
to indicate which two terms are part of the propagator. Here we drop the subscript and assume that Wick-contracted terms are always in the interaction picture
| (162) |
This is the simplest case of the Wick theorem. The more general case is
| (163) |
To calculate time-ordered products like this we need to sum over all possible ways of grouping fields into pairs using the Wick contractions. As an example, let us consider the case of four fields
| (164) | ||||
where we use to save space. We already know how to evaluate Wick-contracted terms using , i.e.
If we are considering vevs the uncontracted terms drop out and we only need to consider all terms contracted, i.e. for example the last line of (164).
Proof of the Wick theorem
We will use a proof by induction. Our base case is which we have already shown. For the induction step we assume the theorem holds for fields and assume that w.l.o.g. the fields are time-ordered, i.e. . We have
| (166) |
where we have applied the Wick theorem for . We now need to move the into the normal ordering. The is trivial because it is already where it is supposed to be. To get the in we need to commute it all the way through the product. For fields that are already involved in a contraction, this is trivial as is just a number and commutes with everything. This means it is sufficient to only consider uncontracted fields. To simplify the notation a bit, we will write down the case without contractions but the generalisation is trivial once a suitable notation is developed
| (167) |
Since commutes with the part of , we have
| (168) |
which is exactly what we wanted to show.
Let us know use what we know to find a graphic representation of these contractions. Consider the vev of (164)
| (169) | ||||
| (170) |
Remember that the are points in spacetime and contains the part of the amplitude that moves a particle from to (or vice versa). We can therefore draw diagrams to represent these terms
| (174) |
This type of diagram is known as a Feynman diagram and they will soon get more interesting. The three diagrams in (174) merely encode all three ways particles can move between the four positions.
Note that if we had three fields we would have found
| (175) |
because all of the terms have an uncontracted field, i.e. they are .
3.4 Asymptotic states and the interacting vacuum
Before we can develop Feynman diagrams further, we need to go to two diversions: one of them relevant, one less so.
The first point is related to the asymptotic states that we defined in (134)
| (134) |
This operator creates a particle at position . However, it is often more useful to think in momentum space and instead create a particle of momentum as we did in Section 2.5. For this, we defined (102) for the free field which we translate to the fields
| (176) |
Applying on this state
| (177) | ||||
| (178) |
where we have used that and . Of course we can repeat the same construction for the out states as well. This is the connection between matrix elements and the vevs of time-ordered products that we have been calculating with the Wick theorem.
The nature of the vacuum
This diversion is not particularly important for the applications of QFT but is quite a fundamental building block of the theory. So far we have been using both for the ground state of the free theory and of the interacting theory. Unfortunately, since the theories are not the same, there is no reason that the two ground states should be the same (or even that there should be a relation between the two). Here, and only here, we will distinguish between the free theory’s vacuum and that of the interacting theory .
We define the energy zero as which means that, in general, the energy of the interacting vacuum will be different . To relate the two Heisenberg states, we apply the full theory’s time evolution operator to the free vacuum
| (179) |
with eigenstates and eigenenergies of the full theory which includes the ground state . Since per definition, the ground state has the lowest energy for all . In the limit all terms vanish since but the ground state’s contribution will vanish the slowest
| (180) |
We can now solve this for and obtain
| (181) |
To make this a bit easier to use, let us shift with and add a factor which, if applied to , will give one
| (182) | ||||
| (183) |
This means that we can obtain the interacting vacuum from the free vacuum by evolving it from the distant past () to the present . A similar construction is possible for where we need to choose to be the opposite sign
| (184) |
We can now write the correlator for
| (185) | ||||
We can simplify things using and assuming the times are properly ordered
| (186) |
Assuming , we can write
| (187) |
to cancel the prefactor phase and
| (188) |
This is completely time-ordered and would have also held for . We can therefore write it as a time-ordered product and use (151)
| (189) |
This result is known as the Gell-Mann and Low theorem and it allows us to slightly formalise what we have been doing (which amounts to ignoring the denominator).
Haag’s theorem
For the above discussion, we have assumed that both and exist in the same space and that their overlap . In practice, this is not true, making the construction invalid. Further, the operator , that we have used to relate the free states to the interacting ones, does not exist either. This result is known as Haag’s theorem and it seriously jeopardises the construction of any QFT. Luckily for us, there are a number of ways to, if not rescue the proof, at least stabilise it enough to be used in calculations. This problem is one of the many issues facing a truly axiomatic construction of QFT.
3.5 Feynman diagrams
To study our first non-trivial Feynman let us consider , i.e. the propagator of the interacting theory. We will assume implicitly that these operators are in the interaction picture even though they are not contracted just yet. For this we use the Gell-Mann-Low theorem (189) but will ignore the denominator for now. We have
| (190) | ||||
| (191) |
Here we have substituted in and combine the integration with the integration. The first term just corresponds to . For the second, we write
| (192) |
We can now apply Wick’s theorem to calculate this vev.
Suggested Exercise
Write down all 15 contractions explicitly to show the following
| (193) |
We can visualise this using Feynman diagrams
| (194) |
We usually define the symmetry factor, i.e. the 3 and 12 to be included in the diagram. In a Feynman diagram like this, we have propagators (the lines of the diagram) and vertices (points where four lines meet). The vertices are located at positions in spacetime that depend either on the process ( and ) or are integrated over (). The number of contractions that contribute can grow quite quickly.
Suggested Exercise
Repeat the above for the term to show that
| (195) |
We can express this using Feynman diagrams as
| (196) |
Here we have written the symmetry factors explicitly even though we previously stated that we consider them part of the diagram. This is for illustration purposes.
Disconnected pieces and the Gell-Mann-Low denominator
In (194) we encountered two types of Feynman diagrams: a connected diagram where all lines and dots are connected; a disconnected diagram where a diagram was free-floating. Let us focus on the disconnected diagram contribution which evaluated to
| (197) |
Once again we find ourselves in a situation where our prediction includes the volume of space time which is obviously infinite. For now let us just call this first disconnected piece , the next more complicated disconnected piece is etc. We will not really worry what these evaluate to because we will soon see that they cancel.
To see what happens, we start with the full series in
| (198) |
Note how in this expression we keep finding the same pieces, both for the connected part and the disconnected ones. We will now try and exploit this structure by rearranging this infinite series by collecting terms not by their power in but by their diagrammatic topology. This will lead to us factoring our the connected pieces
| (199) | ||||
| (200) | ||||
| (201) |
We can do one more step of rearranging
| (202) | ||||
| (203) | ||||
| (204) |
This can now be written in terms of an exponential, summing all disconnected diagrams to all orders in
| (205) |
What we have calculated here is the numerator of the Gell-Mann-Low theorem
| (206) |
We can now consider the denominator which has the same structure but no and . We can use the same logic to show that
| (207) |
Therefore, the disconnected contributions cancel and we are left with
| (208) |
Note that the diagrams we drew in (174) are not disconnected because they are still connected to some external points and therefore do not factor out. This is why the disconnected diagrams are sometimes called vacuum bubbles or vacuum-to-vacuum transitions.
Deriving Feynman diagrams like this is not very efficient. Instead, we usually go the other way around and draw all possible diagrams and then use Feynman rules in position space
| For each internal line | (209a) | ||||
| For each vertex | (209b) | ||||
| For each external line | (209c) | ||||
| Divide by the symmetry factor. | (209d) | ||||
We obtain a factor from the Taylor expansion which cancels with the ways of arranging the vertices. Further, there are ways to arrange the lines going into a vertex which cancels with the in so that our vertex rule is just . After these factors are accounted for we usually still overcounted the diagram. To avoid this, we add the diagram’s symmetry factor which are for example explicitly written in (196) Formally, the symmetry factor is the order of the symmetry group of the diagram that keeps the external lines fixed. A more practical set of rules that will cover almost all use cases is
-
•
lines that start and end in the same vertex, add a factor of 2
-
•
propagators connecting the same two vertices, add a factor of
-
•
if two vertices are equivalent, add another factor of 2
The rules as formulated above are valid in position space. Often it is more suitable to have them in momentum space where we write the propagator as a Fourier transform (cf. (123))
| (210) |
We now assign a momentum to a propagator and split the factor into both ends of the line. This means that for internal vertices we now have
| (211) |
In other words, momentum is conserved at each vertex. With this, we now have our momentum-space Feynman rules
| For each internal line | (212a) | ||||
| For each vertex | (212b) | ||||
| For each external line | (212c) | ||||
| Impose momentum conservation, | (212d) | ||||
| Integrate over unconstrained momenta | (212e) | ||||
| Divide by the symmetry factor. | (212f) | ||||
Suggested Exercise
Calculate the symmetry factors of the following diagrams
| (213) | |||||
| (214) | |||||
| (215) | |||||
| (216) | |||||
| (217) |
Suggested Exercise
Write down all diagrams that contribute to the four-point function up to and calculate the amplitude for .
At we have a single diagram
| (218) |
At , we have four interesting diagrams
| (219) | ||||
| (220) | ||||
| (221) |
There are also diagrams where a correction is applied to the external line
| (222) |
We will discuss these momentarily.
Suggested Exercise
Consider the theory
| (223) |
Calculate the symmetry factor of the following diagram through the Wick theorem
| (224) |
Convince yourself that its Feynman rules are the same as in except for the vertex
| For each vertex | (225) |
Calculate up to .
3.6 Returning to the matrix
To use our formalism to calculate matrix elements we would have to repeat the derivation of the Gell-Mann-Low theorem for states other than . However, we had to use the fact that the vacuum is the state with the lowest energy which will not be true for any state that contains particles we would like to scatter. It is possible to rescue this argument but how to do this goes well beyond the scope of this course. Instead, the following construction should motivate why we might think that we can calculate matrix elements using Feynman diagrams.
Consider the matrix element (cf. (142)) between an outgoing state (composed of particles with momenta ) and an incoming state (composed of particles with momenta )
| (226) |
The indicates that we have ignored the equivalent denominator of the Gell-Mann-Low formula (189) in that hopes that it will cancel again if we only consider connected diagrams. Since the external states can be expressed through field operators and the vacuum, we can make them part of the Wick contractions that defined the Feynman diagrams.
However, this will lead to plenty of diagrams where nothing of interest happens like the ones in (174). To avoid this, we define the interesting part of the matrix
| (227) |
where contains only the connected diagrams and the all the cases where no interaction takes place. Since our Feynman rules imply momentum conservation, we have made this explicit already here66 6 Sometimes you will see this factor to be defined as part of the matrix instead..
There is one more restriction on the types of diagrams that enter . Consider the following diagram which is fully connected
| (228) |
with the intermediary momentum . Momentum conservation forces and because we want to be on-shell, i.e. , we now have a singularity. This is quite a big problem. will only make sense if we exclude this type of diagram where a loop is attached to an external leg like this. One can show that these types of diagrams are similar to the vacuum bubbles we have already excluded in (208). Diagrams that do not have this problem are called amputated.
Let us summarise our achievement
| (229) |
There is one final subtly related to this that we will revisit later in Section A.