Warm Quantum Fields

What happens when there is more than a couple of particles in space? We will use thermodynamics to introduce perturbative corrections to QFT calculations. We start with a revision of the basic structures in statistical physics, formulated in a more abstract way such that it is directly applicable to the path integral formalism. Then play with the partition function for QFTs enough to derive corrections to the Green's functions of the theory and finally do some perturbative corrections to show that the temperature is meaningless and we were right to ignore it all along.

These notes rely on:

  1. Kapusta - Finite temperature field theory
  2. Bellac - Thermal Field Theory
  3. Ramond - Field Theory
  4. Donoghue - The analytic continuation of distributions
  5. Peskin - Introduction to Quantum Field Theory
  6. Tong - QED
  7. Weldon - Covariant Calculations in Finite Temperature Field Theory
  8. Ahmed - Vacuum Polarization at Finite Temperature in QED

Ensembles

The playing field of statistical physics is ensembles which is a term that comes from probability. The idea is that one has a measure space that denotes all the possible outcomes for the big system and then they add a probability measure that satisfies certain physical conditions such as the existence of temperature or chemical potential and so on.

The definition of an Ensemble is useless because it is too general, but here it is but that's too abstract to make sense.

Definition: An ensemble (Γ,Σ,λ,S) is a measurable space Γ with sigma algebra Σ and a map λ:S×Σ[0,1], where S is some set and for any sS λs=λ(s,) is a measure on (Γ,Σ).

The definition of the other components is here and honestly it isn't even that relevant. An ensemble is basically a space Γ that contains all the configurations of the constituents of the system and a probability measure λ slapped on top of that such that we can tell how likely some configurations are versus others.

The physics of the ensemble are all inside λ. Usually it depends on the Hamiltonian in a cute way, or the chemical potential, temperature, and so on. These are the parameters in the set S. Usually S is something like R representing time, where we get a different measure for every point in time. Other times S can be the set of all temperatures or chemical potentials or both. Let's see how that works.

Quantum Mechanical Canonical Ensemble

It is actually worth to go through this as an example because it will force us to think about interesting topics such as the spectral theorem and so on, plus it will be a useful motivating intuition for our subsequent definitions.

Example: (Quantum Canonical Ensemble) Consider a quantum system with Hilbert space H that describes the states for a single particle. Then we can consider the Fock space of that system which is the Hilbert space that contains all the possible combinations of this kind of particle and it is given by

F=n=0Hn=CH(HH).

In any physical theory one doesn't just get the Fock space by itself, but also gets an algebra of possibly bounded self adjoint operators A, or observables, that contains familiar physics concepts such as the Hamiltonian H, the number operator N, etc. Honestly there is a generalization for algebras of unbounded operators but I don't know it, and I don't need to. It exists.

Given all that, we need to convert F into a measure space by introducing a sigma algebra. We are lucky because there is a free sigma algebra in the projective Hilbert space given to us by the Born rule!

Definition: The projective Hilbert space P(H) is the set of all rays of a Hilbert space H, namely

P(H)={[ψ]H{0}ϕψλC s.t. ψ=λϕ}.

Corollary: The projective Hilbert space is isomorphic to the set of all projection operators to the lines in H.

The projective space has a natural Khaler manifold structure with metric induced by the inner product in H. Here is how.

Theorem: Consider a complex Hilbert space H with the canonical metric induced by its hermitian form

g=dzμdzμ,

where zμ are projection coordinates. Then P(H) can be identified as the quotient of the unit sphere SHH on the Hilbert space, and the induced metric, known as the Fubini metric, makes it a Khaler manifold.

This is way too complicated for what it is. It just means that the probability norm we are using for the born rule defines a metric on the projective space and more importantly a volume form there, including a measure.

Example: Let ψ,ϕH, then their geodesic distance in R(H) induced by the Fubini metric is given by

γ(ψ,ϕ)=arccos|ψ,ϕ|ψ,ψϕ,ϕ.

This can be thought of as their angle in the projective space which belongs in [0,π/2].

We actually don't care to measure F we only really care on measuring P(F) which we can do with the Fubini metric. However, there is one super awesome theorem that can define the measures so much simpler. Here it is.

Theorem: (Gleason's Theorem) I probability measure is a Fubini probability measure μ on P(H) iff there exists a unique positive semidefinite linear map ρ:HH on H with unit trace Trρ=1 such that the measures density is given by the map

p:P(H)[0,1]ΠTrρΠ,

where Π:HH is an element of P(H) thought of as a projection operator to a line in H.

Note that the integral of the Fubini form on P(H) is normalized to 1 for this to work.

Intuition: In physics we call ρ a density matrix. The above theorem can be rephrased to any density matrix defines a distribution over the projective space.

In particular, the configurations in quantum mechanics are really only the elements projective space of the Hilbert space. This is the rigorous form of the statement: "Physical states are equivalent up to a phase." As a result, the density states are essentially a probability distribution over the set of such configurations.

In a more quantum mechanical language a density matrix can be thought of as the probability distribution that a quantum system would collapse to a specific state after observation. For example, the density matrix |00| in a spin-12 system is a probability distribution over all the possible quantum states that after observation the system would collapse to the |0 state.

This is similar to classical statistical mechanics, where the density of states is a distribution over the classical configurations.

So now we can add measures using density matrices which is honestly pretty cool! We are ready to define an ensemble.

Definition: Consider a quantum system with Hilbert space H, a Hermitian map H:FF (i.e. the Hamiltonian), and a Hermitian map N:FF such that for any ψHn Nψ=nψ (i.e. the number operator). The quantum mechanical grand canonical ensemble is an ensemble (P(F),Σ,μ) where (P(F),Σ) is the Fubini measure space on the projective space, and λ:R2×Σ[0,1] is a collection of measures given by

(β,μ,A)λβ,μ(A)=1Z(β,μ)ATr(eβH+μNΠ),

where is the Hodge star on P(F) with the Fubini metric and Z:R2R is a normalization factor called the partition function and given by

Z(β,μ)=P(F)Tr(eβH+μNΠ)=Tr(eβH+μNΠ),

where the last equality is given by the spectral theorem.

And here we have it! A fully fledged ensemble! Probably not the best first example for this, but this can serve as a blueprint for the later definitions.

Physical Motivation

Notice that the grand canonical ensemble is the ensemble given by the density matrix

ρ=eβH+μNZ.

Let's go over the physical motivation for this. The grand canonical ensemble is supposed to describe a system where there are a bunch of particles at some fixed temperature that is connected to a reservoir such that adding or removing particles expends a fixed amount of energy. The total energy remains constant though. Let's motivate each condition separately.

Such a system thermodynamically has to obey the first law of thermodynamics which in this case is written by

dE=TdSμdN,

where we define the number of microscopic states that correspond to a particular macroscopic state to be

Ω=eS,

by where S is the entropy which is some function over the macroscopic configuration space. The probability is going to be proportional to Ω since it indicates the number of available states to our system with given macroscopic variables. Now consider the system attached to a big reservoir. This corresponds to taking the (categorical tensor) product of the two configuration spaces, so the total number of states ΩT is going to be ΩSΩR, i.e. the product of the available states of our system S and the reservoir R. This implies that

ΩT=ΩSΩReST=eSS+SRST=SS+SR.

Now assume that the reservoir is much bigger than our system. This really means that if we change its state variables the reservoir won't budge. To be more precise, we focus on the Reservoir. It is a system whose total energy is given by

ER=ETH,

where H is the Hamiltonian of our system. The issue is that the system is not in equilibrium unless H=0. Therefore our thermodynamics only work for that regime. Here is where the large assumption comes in. The picture below is the intuition of what it means for the two energies to be close. The bottom blob is the surface in the macroscopic configuration space where the total energy is ER, while the other surface is the one with the Hamiltonian.

macroscopic-phase-space

What the condition means is that these surfaces are physically close in the configuration space. In other words, for every path γ from the bottom surface to the top, defined by some evolution of the Hamiltonian any observable O can be written as

Oγ(ϵ)=OR+Oϵϵ+O(ϵ2),

where ϵ is some small unitless parameter. In addition to that, we have that at any point in the configuration space

TdSR=μdNR+dER.

Therefore we can now calculate the difference in entropy. This directly leads to the weighting factor we have introduced above.

Partition Function

Almost by accident we got introduced to the partition function in the definition of the grand canonical ensemble above as this normalization constant. It very often happens to be a smooth function and it can give us information about all thermodynamic observables by taking appropriate derivatives of it. In some sense, one has an Ensemble simply by knowing the partition function.

To see how that is true we will first define the average of observables in our thermodynamical system.

Definition: Let (Γ,Σ,λ,S) be an ensemble and f:ΓX be a map from the configuration space to some Banach space X. Then the ensemble average of f is a map f:SX given for any sS by

f(s)=Γfd(λ(s)).

where the integral is the Bochner integral (generalization of Lebesgue integral for Banach spaces).

With this we can see that the partition function is very close to the ensemble average. In particular the partition function helps us define a probability measure using a volume integral.

Proposition: Let μ be a measure on (Γ,Σ) then the measure λ:ΣR¯ defined for all AΣ by

λ(A)=μ(A)Z,

where Z=μ(Γ) is the partition function, is a probability measure.

Proof: Clearly λ(Γ)=μ(Γ)μ(Γ)=1 for finite Z. The proof also extends to the other case using appropriate limits.

So in some sense, the partition function can help us relate one measure as a function of another, but it can also help us obtain ensemble averages. Specifically check out this example.

Example: (Cool stuff on manifolds) Let Γ be an orientable Riemannian manifold and consider it as a measure space using integration of top forms. Now we can define a collection of measures μ:SC(Γ) parameterized by another manifold S. Given such a collection of measures, one can define a partition function Z:SR obtained for any sS by

Z(s)=Γμ(s).

Then one can define the collections of probability measures λ:SC(Γ,I) by λ=μ/Z. Then we have that for any vector XsTsS for some sS

Xμμ=XlogZ.

The proof for this is through the following calculation

XlogZ=XZZ=1ZΓXμ=ΓXμμμZ=ΓXμμλ.

This is a pretty powerful fact, because usually in thermodynamics S is some set of thermodynamic parameters μ is an exponential function that includes all of our interested quantities like the Hamiltonian, number of particles, etc.

Given such a successful construct we would like to use it in our path integral formalism.

Partition Functions in QFT

As we have seen in the previous section a partition function doesn't exist without a reference to a measure. In quantum field theory the configuration space is a projective space of a Fock space that is comprised out of all the field operators. There we have introduced a measure called the path integral that allows us to integrate over everything and assign probabilities to each state just like in the quantum mechanical grand canonical ensemble we saw in the previous example. We want to find the partition function with respect to that measure in such a way that it encodes the idea of temperature.

Aside on Complexification

Let's begin by considering what is the condition under which one can continue a real valued distribution to a complex analytic function in some sense. Let's first define the object that would make sense in this context.

Definition: Let UC and ΩR then an analytic distribution valued function is a map T:UD(Ω) such that for any ϕD(Ω) the complex valued function T(ϕ):UC is analytic.

Recall that D(Ω) contains test functions of the form ϕ:ΩC. The relevant theorem is as follows.

Theorem: Let VC be an open subset of the complex plane such that UV is a proper subset. Additionally, consider an analytic distribution valued function T:UD(Ω) for some measure space Ω. Then there exists a unique analytic valued function T:VD(Ω) such that for any test function ϕD(Ω) and point zU

Tz(ϕ)=Tz(ϕ),

if and only if for all such test functions T(ϕ):UΩ can be analytically continued to V.

This result is what facilitates Wick rotation in field theory. The fields are such distribution valued functions usually with discrete singular support.

Transition Amplitudes

In Quantum field theory we want to calculate things like the amplitude that a state, after time evolution, will evolve to a different state. These are called transition amplitudes and can be calculated with the path integral prescription that we briefly describe here.

To make this more precise consider the eigenstates of the field operator ϕ(x,0) defined by

ϕ(x,0)|α=α(x)|α,

where α is some section of the appropriately chosen time slice. Following a lattice limiting procedure one can derive that the amplitude between two field eigenstates is given by

α|eiHt|β=ΩαβDϕi0tL(ϕ)dt

where the domain Ωαβ is populated by the functions with appropriate boundary conditions like so

Ωαβ:={ϕΓ(E)ι0ϕ=β,ιtϕ=α},

where ιt:ΣtM is the inclusion map of the corresponding time slice. The partition function looks extremely similar to this! So let's try it.

Free Theory Partition Functions (Intuition)

These path integrals are really only well defined in free theories. So let's try to do one derivation and then we can move on to the interacting ones.

We have seen from the previous example of the quantum mechanical phase space that the appropriate partition function is given by

Z=TreβH,

where βR is some real number related to the temperature. We could create a field theoretic equivalent of this statement by calculating the trace in the basis of eigenstates of the field operators. Without loss of generality we will do this for a single field since theories with multiple fields can be written as single field theories.

Specifically, we would have (up to an overall factor that we divide out in all calculations)

Z=dαα|eβH|α,

where the integration is carried out over all field configurations. The cool thing is that we are already familiar with such constructions because we can Wick rotate. In other words, we analytically continue our fields to the complex plane such that plugging in τ=it in the fields is a sensible thing to do. Once we do so, the action changes because we are integrating over the imaginary axis instead. Specifically, we obtain

Z=dαΩααe0βLEdτ.

Now is where things get suuuuper cool. What we have here is that we are integrating the domain Ωαα over all configurations α. One such element of Ωαα is one that starts at τ=0 at α and ends at α at τ=β. Therefore this is sort of like a field on a circle. Here is a picture.

Compactification-picture

In this picture we have a circle with perimeter β parameterized by arc-length τ on top of which two fields satisfying the boundary conditions are depicted in blue. As a result we can notice that integrating over all the possible α we are summing over all the possible field configurations on this circle without double counting (since each set of fields differs at at least one point). Therefore the partition function for equilibrium temperature β can be equivalently written as this trace over a field theory on a circle with perimeter β. This yields,

Z=DϕeS1LE(ϕ),

where the integral is taken over all the fields in this new compactified configuration. With this motivation we are ready to rigorously treat the free field theoretic partition function.

Free Bosonic Theory Partition Functions

We want to consider a free bosonic theory and calculate its partition function. To do this we assume that underlying spacetime is some Riemannian manifold M=Rn1×S1 and the fields are sections of a vector bundle π:EM with rank k, equipped with a Euclidean (or Hermitian) bundle metric, that has a cute connection :X(M)×Γ(E)Γ(E) so that the fields can move nicely. The Euclidean Lagrangian LE:Γ(E)Ωn(M) is then given for any fields ϕΓ(E) by

LE(ϕ)=12dϕdϕm2ϕ,ϕ,

where the wedge product incorporates the bundle metric on the fibers of E. Now we can define the action S:Γ(E)R as

S(ϕ)=MLE(ϕ).

We also pick a metric on M such that the volume of the circle is βR. Using integration by parts we can write the action as

S(ϕ)=12Mϕ,(dd+m2)ϕ=12ϕ,DϕL2,

where D is a linear differential operator on Γ(E) and the second inner product is the inner product in L2(E). Therefore the partition function can be written as

Z=Dϕe12ϕ,DϕL2(detD)12,

where we have used some functional integration techniques from Peskin. To evaluate this determinant we need to find the eigenvalues of D. Luckily since our space is quite flat it would be possible to do this let's express ϕ in Fourier modes.

Lemma: Let ϕ:S1V be a smooth function to a finite dimensional vector space V over R. Then there exist {ϕn}nZV and {ωn}nZR such that for any tS1

ϕ(t)=nZϕnei2πnt.

Now we are ready to do this! Now, however, we have a different metric in our circle such that the integration measure yields

S1dτ=β,

instead of 2π. Using this different metric, the expansion of any field now reads

ϕ(t)=β12nZϕneiωnt,

where ωn=2πnβ. We usually call these frequencies Matsubara frequencies.

This has the following interesting consequence.

Corollary: Any field ϕL2(E) can be written for any point (x,τ)M as

ϕ(x,τ)=β12nZRn1dn1peipx+iωnτϕn(p).

Notice that ϕn(p)=ϕn(p) in order for ϕ to be real valued. If the connection on E is trivial, ϕn(p)=ϕnm(p)em where em=0.

The really interesting thing is that we are rarely interested in the partition function itself, but rather in its log and its derivatives. Therefore we only need to calculate the log of the determinant of D which is its trace! From now on we will assume that the connection on E is flat. It also works for non flat connections, but we would have to sum these eigenvalues, etc. which is getting too much for this example.

Proposition: The logarithm of the partition function is given by

logZ=12TrlogD.

up to some real number AR.

Proof: We use the fact that deteA=eTrA.

The trace is much easier to calculate than the determinant, especially since we have found a basis for all the fields given by

{eixp+iωnτemnZ,pR3,1mk},

where {em}1kV is the standard basis of the fiber. But more importantly this is not just a basis, it is an eigenbasis! So much luck and love came out of this choice. Here is why.

Corollary: The trace of logD is given by

TrlogD=nZRn1dn1plog[β2(ωn2+ω(p)2)],

where ω(p)=p2+m2.

Proof: First let's see what happens when we calculate

Deixp+iωnτem=β2(2τ2Δ+m2)eixp+iωnτem=β2(ωn2+p2+m2)eixp+iωnτem.

if we define ω=p2+m2 as usual, the eigenvalues are

λn(p)=β2[ωn2+ω(p)2].

The reason why this is nice is that for an eigenvector ϕ of a linear operator A=elogA we have that

Aϕ=λϕ=elogλϕ=elogAϕ.

Therefore the trace of the log, its going to be the sum of the eigenvalues which is given by

TrlogD=knZRn1dn1plogλn(p).

With this we have proven the statement.

Ok, but while this calculation was satisfying it is still not interpretable in a nice way. In particular, we would like to get some thermodynamics out of this about the particles of the theory in some ensemble. So let's start noticing some things.

Lemma: The following identity holds

S(x):=nZ1n2+x2=πcothπxx.

Proof: Consider the identity

f(x)=sinhπxπx=n=1(1+x2n2).

Then we can take

logf=n=1log(1+x2n2).

By differentiating we deduce that

ff=2xn=11n2+x2=2x(12S(x)12x2).

On the other hand

ff=πcothπx1x.

Equating we obtain the identity.

In addition to this silly proposition we also need to remember that the partition function is the same up to a constant factor (i.e. independent of β). From now on we will denote that equivalence . Therefore we can write the partition function is equivalent to

logZk2nZRn1dn1plog(n2+α2)k2nZRn1dn1p1αdξ2n2+ξ2,

where α=βω2π. Notice that inside this integral we encountered the sum of the previous lemma. Therefore the partition function is equivalent to

logZπkRn1dn1p1αdξcothπξkRn1dn1plog|sinhβω2|.

where we have been dropping terms independent of β along the way. This is a cool result because we can reinterpret it to

logZRn1dn1p[kβω2klog(1eβω)].

which is the partition function for Bose Einstein! Yey. This includes the energy of the vacuum so we can subtract it because by definition we pick the vacuum to be at 0 energy. But still.

Theories with Interactions

Now that we have some calculations under our belt, it is time to play around with theories where interactions can happen. Such theories have action

S=S0+SI,

where S0 is the action of some free theory, and SI is the action that contains only interaction terms in the fields. Therefore we can write down the partition function as

Z=DϕeS0+SI=DϕeS0n=0SInn!.

Since we are not interested in the partition function itself, but its logarithm we can write

logZ=log[DϕeS0]+log[1+n=01n!DϕeS0SInDϕeS0]=logZ0+logZI,

where Z0 is the partition function of the free theory, and ZI will be the correction due to interactions. The reason for expressing it this way, is because the quantity inside the sum is what we actually know what to compute, which is computing the interaction terms in the background. For that we can use diagrams. We will explore this in detail in the following example.

The Generating Functional

We often use the partition function as a generating functional for the correlators, so it would make sense for us to do it here too. Here is how to extend that definition.

Z(β,J)=DϕeM¯LE(ϕ)+Jϕ=Z(β,0)e12M¯2GJ×J

where M¯Rn1×S1 is the Euclidean space with compactified time an\phi(x')d G is the Matsubara propagator. Notice that Z(β,0) is the partition function we have found before. Using this tool we can derive generalizations of Wick's theorem and time ordered correlators for the finite temperature expectation values of observables given by the Born rule.

The Matsubara propagator satisfies the equation

DG=δ,

where D is the differential operator that we have found before.

First order corrections to the propagator

It is clear that one can obtain the propagator using the generating functional above. Now we can calculate its first order correction in the coupling, using Wick's theorem.

The 2-point correlation function is given by

ϕ(x)ϕ(y)=Dϕϕ(x)ϕ(y)eM¯LE(ϕ)DϕeM¯LE(ϕ)=δ2logZδJ(y)δJ(x)|J=0.

To first order we can approximate this using perturbation theory on a well defined potential term. However, we often want to calculate such expressions in momentum space (in order to define things like polarization etc.). This results in ugly expressions involving sums over all the Matsubara frequencies so we would really like to figure out a way to do this in "real-time."

Real Time Formulation

For real time we have found all our expressions in the compactification of the Euclidean space that we got access to through Wick rotation. What we want to try now is to motivate how the formalism exists back in the Minkowski space. We will start by motivating this through the definition of Feynman rules.

Feynman Rules for T0

Consider a Lagrangian with interaction term V(ϕ). Therefore the generating functional can be written as

Z(β,J)eM¯V(δδJ)e12M¯2GJ×J.

This is exactly the same generating functional as the one we had for 0 temperature with two exceptions. The first is that the propagator is now the Matsubara propagator on the compactified Euclidean space M¯, and the second is that we are integrating over the compactified Euclidean space. Yet, we can symbolically derive Feynman rules without any difference for the procedure at T=0. The differences are

  1. The integration measure in momentum space becomes proportional to

    TnZRkdkp(2π)k.
  2. And the vertex lines get delta functions on the momenta and positions which implements "momentum conservation" but with different normalization.

Complex Time ordering

One could now notice that the measure in momentum space could be written (with some appropriate meromorphic functions) as a integral over an carefully chosen contour over the full (not compactified) space. Therefore we could use that to move to the real time picture via introducing a consistent notion of Time ordering.

Definition: A thermal Green's function ordered on a contour γ:RC parameterized in a monotonically increasing way is a distribution Gγ given by

Gγ(x1,x2,,xk)=Tγϕ(x1)ϕ(x2)ϕ(xk)β,

where Tγ is the γ-ordering operation defined as follows.

TγA(x)B(x)=θ(γ1(t)γ1(t))Aγ(x)Bγ(x)+θ(γ1(t)γ1(t))Bγ(x)Aγ(x),

where θ:R[0,1] is the Heaviside step function, and the notation Aγ(x)=A(γ(t),xi) stands for the analytic continuation of A along γ. For a complete definition, given a countable set of operators A={Ai(xi)}iI indexed by a partially ordered set I there exists a permutation σ:II such that for i,jI then

γ1(ti)γ1(tj)σ(i)σ(j).

The the γ-ordered product of the operators in A is

TγA(xi)=iIAσ(i)(xσ(i)).

The thermalness of the propagator depends on the path we choose we will soon find conditions on what this path should be, but for now let's try and see what implications does this have on the propagators.

Propagators

Not any curve γ would correspond to our intuition of thermal, but we can naively see that if it does when we wick rotate the fields we are considering must fall inside [0,β] therefore it suffices to consider curves such that Imγ(t)[β,0] (since τ=it). Therefore the curves are actually segments in this domain.

Therefore we can consider curves of the form γ:[T,T]C where TR is some time that we will later take to infinity.

Additionally we want the 2-point functions to be defined by

Gγ(x,x)=Tγϕ(x)ϕ(x)β=θγ(tt)ϕγ(x)ϕγ(x)β+θγ(tt)ϕγ(x)ϕγ(x)β,

where θγ:=θγ1. The interesting thing is though that we can actually calculate these correlations in the Matsubara formalism to derive the following property.

Theorem: The following is true regarding the 2-point thermal correlation function

ϕ(x)ϕ(x)β=ϕ(x)ϕ(x+iβ)β,

where the addition is understood to apply at the time component, i.e. x+iβ=(x0+iβ,xi).

Proof: To show this we need to expand things slightly, namely

ϕ(x)ϕ(x+iβ)β=1ZTr[eβHϕ(x)eβHϕ(x)eβH]=1ZTr[eβHϕ(x)ϕ(x)]=ϕ(x)ϕ(x)β.

By the way, this implies the need for periodic boundary conditions on the time axis but in a slightly more general way that applies to the real time formalism also. Now we are ready to continue with our derivation of the propagators.

Consider the Fourier transform of the correlation function

Dβ>(xx):=ϕ(x)ϕ(x)β=:Dβ<(xx).

The Fourier transforms (using translation invariance) have the property that

Dβ<(k0,k)=Dβ>(k0,k)=eβk0Dβ>(k0,k).

This results in the fact that we can write the propagator as

Dβ>(k)=(1f(k0))ρ(k)Dβ<(k)=f(k0)ρ(k)

where

f(x)=1eβx1,

and ρ is some undetermined envelope function called the spectral function.

Proposition: The following identity holds

f(x)+f(x)=1.

Therefore the propagator can finally be written like so.

Lemma: The thermal propagator on path γ can be written as

Gγ(xx)=R4d4k(2π)4[θγ(tt)+f(k0)]ρ(k).

Proof: This is done by writing the propagator in terms of Dβ>,< and then using the following proposition.

This is an awesome result because it simplifies so many things! For example, we can derive the free propagator by taking

ρ(k)=2πϵ(k0)δ(k2m2).

Contours for the Real Time Partition Function

One of the weird results is that the real time propagator is not the analytic continuation of the imaginary time (Matsubara) propagator. Here is how to calculate what it actually should be.

What we will derive is that we need to double the degrees of freedom of our theory in order to obtain the real time propagator. To do this we will find the path γ that corresponds to the thermal temperature.

Essentially the 'thermal' condition is the thing that we derived earlier. A field is thermal if it is iβ periodic in time. Taking the path integral with respect to such fields will naturally give the trace of eβH. We did this already and calculated that the periodic wick rotated fields are such fields. This required us to calculate all our observables in imaginary time, but the natural question to ask is if it is possible to calculate our observables in real time.

Let's go back to the partition function which we calculated by the trace

Z=TreβH=dαα|eβH|α.

What we secretly did here is that we said that the integral over the field in "real time" that appears in the action, is the same as the integral over the fields in "imaginary time" because the action is not expected to have any singularities in the contour below.

Wick-Contour

So what we said was that as long as no singularities are present on the two quadrants, the total action on a field analytically continued to that contour will be zero. Therefore we can split the action into contributions from the field on the real part (which is the action that we have started with) and contributions from the imaginary part, which is the action we are using in the "imaginary time" formalism. Then we applied the boundary condition by the trace to obtain functions that are defined on the following contour.

Wick-Contour-Compactified

In this picture we have drawn the semicircular part of the contour that we take to be asymptotically far away with green. This is to illustrate that the action of fields defined on this contour vanishes as the contour is taken asymptotically far away.

Ok, but so far we haven't actually introduced an equivalent thermal field that lives on the real line! To do this, we can introduce the following contour on the complex plane and then compactify, i.e. add the trace boundary condition, to see what's up.

Real Time ContourCompactified Real Time Contour
Real-time-Wick-ContourReal-time-Wick-Contour-Compactified

In the compactified picture I removed the lines that were drawn in Green before because this would have been too frustrating to visualize. Here is how we construct the new contour. We pick a time t0R. Then we move by T>0. Then go down by iβ, the left by T, and finally start the previous Wick contour but at t0iβ instead of 0. Now the action ends up having multiple contributions from fields living on each part of the contour. The claim (which is true for most insertions in the path integral) is that the contributions to the action by the blue and green imaginary time fields exactly cancel when one takes the path integral.

While this cancellation doesn't always happen there are ways to take limits of this contour such that it does. One such limit, for example, would be to move the two parallel red parts arbitrarily close to each other. Yet after such cancellation what we are left with is the contributions to the action by two fields that live on the red parts! This is awesome!

By definition if there are no singularities enclosed, the contributions by fields on the two contours should not vanish, but the price we paid is that we have introduced two copies of the system. Therefore the real time, thermal action doubles the degrees of freedom of our theory in order to calculate the partition function by "integrating in" a copy of the system.

Relating to the Propagators

The effective result of this, is that the thermodynamic corrections will yield mixing terms in the propagator between the two fields. This sucks a little, but often we can diagonalize the two fields to obtain the real time propagators.

If we consider γ1, γ2 to be the parallel red lines on the picture above we can define Ji:=Jγi for i=1,2. Which gives us the two fields that we are interested in. Putting everything together, we can write down a generating functional with the following property

Z(J1,J2)=Z(0,0)exp[12R4dxR4dyJa(x)GabF(xy)Jb(y)],

where GabF is the free propagator in the new thermal theory with the double degrees of freedom which is defined by

GF=(DF(γ1(t)γ1(t))DF(γ1(t)γ2(t))DF(γ2(t)γ1(t))DF(γ2(t)γ2(t))),

where DF is the free theory propagator that we already know. It is often possible to diagonalize this matrix and what we will get in momentum space is

GF(k)=U1(k)(DF(k)00DF(k))U(k),

where the orthogonal matrix U is given by

U(k)=(1+f(|k0|)f(|k0|)f(|k0|)1+f(|k0|)),

where we derived f(x) before as the Bose-Einstein distribution. This process is very similar to introducing Ghost fields. We in fact call the secondary fields Ghosts.

Self Energy

The self energy can be calculated with its usual definition in momentum space

G1GF1=iΠ,

but now this is a more complicated matrix equation. Nevertheless, using the diagonalization matrix U we can get results that show that Π is also diagonal. Yet there is a further simplification one can make, by relating the propagator components to each other assuming thermal equilibrium.

Finite Temperature Gauge Theory

The whole point of this introduction was to introduce finite temperature Gauge theories. The reason is that gauge fixing at finite temperatures is less trivial than it was at T=0. Let's see why through a calculation of black body radiation.

Incorrect Free Energy

Since we didn't need ghosts for quantizing Abelian Yang Mills let's try and do this thermally and see if it this is still valid. Essentially all we needed to do was to quantize a gauge theory by picking a gauge.

Let's do everything in the Feynman where we have picked this

dA=0.

and we impose the equations of motion coming from the Lagrangian

L(A)=12FF12(dA)(dA).

Now we can mode expand A and the equivalent momentum π to

Aμ(x)=d3k(2π)312ωkλ=03[ϵμ(λ)(k)ak(λ)eikx+ϵμ(λ)(k)ak(λ)eikx]πμ(x)=id3k(2π)3ωk2λ=03[ϵμ(λ)(k)ak(λ)eikxϵμ(λ)(k)ak(λ)eikx].

where there are 4 polarization vectors. Notice that the constraint equation,

ψ|dA|ϕ=0

for any physical states |ϕ,|ψ, does not constrain the polarizations in this case.

However, if we do the same thing in some physical gauge like

A(X)=0

for some vector field X we would obtain something else. Let's pick X=t to go to Coulomb gauge where the field operators and their momenta are given by

Ai(x)=d3k(2π)312ωkλ=12[ϵi(λ)(k)ak(λ)eikx+ϵi(λ)(k)ak(λ)eikx]πi(x)=id3k(2π)3ωk2λ=12[ϵi(λ)(k)ak(λ)eikxϵi(λ)(k)ak(λ)eikx].

where we can see that the polarization vectors are only two!

While this doesn't matter at T=0 when calculating the partition function with one quantization scheme vs the other we will get twice as much free energy!

Specifically the free energy is given by

F=1βlogZ=1βlogTreβH.

In principle we can write the Hamiltonian as the sum of identical single mode Hamiltonians which will result (by permutation and re-parameterization symmetry of the trace) to an overall power on the partition function. In terms of the free energy, which gives directly the black body spectrum, that becomes a factor in front.

One theory has 2 polarizations while the other has 4! Where is the mistake?

The mistakes comes in the fact that while all the operator statements are true the Hilbert spaces we are taking the trace over are different. In most cases in QED we don't actually calculate such traces therefore this never becomes an issue. However, we already know how to fix this issue! The answer is to introduce Ghosts!

Rescuing the Free Energy with Ghosts

We introduce Fedeev-Popov ghosts to properly quantize the electric field in a covariant way. Without delving too much in the formalism we will introduce the following identity.

Theorem: Let F:AA be an endomorphism of a Banach space. The the following identity holds in the path integral over A.

ADαδ(F(α))detJF=1,

where JF is the Jacobian of F which we precisely define later.

In physics we often call F the gauge fixing function because we design it in such a way such that its kernel contains only the fields in a specific gauge.

Definition: Given a gauge fixing function F the M4 Yang Mills partition function can be written as

Z=D4Aδ(F(A))detJFeSE.

Where SE is the Euclidean effective action that includes the gauge fixing term 1ρdAdA.

Usually the determinant of the Jacobian is quite a pain to calculate in practice, however there is a lemma that can help us make it easier. To derive it we need a precise definition of the Jacobian.

Definition: The Jacobian of a gauge fixing function F is the variational differential with respect to the action of a lie algebra element of the gauge group. In particular let ωg be a lie algebra element, then the Jacobian is given by

detJF(A)=det[δω(y)F(A(x))]=det[FaAμc(x)Dμcb(x)δ(xy)],

where D is the covariant derivative.

Lemma: The following identity holds

detJF(A)=DηDη¯eM¯η¯aFaAμc(x)Dμcb(x)ηb,

for Grassmann fields η,η¯, where the integral is the Berezin integral.

These are the Ghost fields! By integrating them in to our partition function we guarantee that the gauge is fixed, and we no longer have to worry about miscounting degrees of freedom or whatever. Let's see this in practice.

Working in the Feynman axial gauge we have that

F=A(3),

carrying out the calculation we have that the effective Lagrangian after the introduction of the Ghost fermions C,C¯ is given by

Leff=12FF+C¯3C=(14FμνFμν+C¯3C).

Up to a total derivative dω this Lagrangian can be written as

Leff=12ADA+C¯3C+dω.

where D=dd. Therefore, we can calculate directly the partition function using the path integral to be

Z=(detD)12det3.

Therefore the free energy is given by

logZ=Tr[12logD+log3].

picking the same basis that we have found before for the compactified space we can diagonalize and with a bunch of effort derive

logZ=Tr[12log(β6p32(ωn2+p2))+logβp3].

Now for the other gauge, we have that F(A)=dA so the effective Lagrangian is (up to a total derivative dω)

Leff=12FF12dA(dA)+dC¯dC=12ADA+C¯ddC,

where D=dd. Therefore we have that the partition function is given by

Z=(detD)12det(dd),

which results to

logZ=Tr[4(12)log(β2(ωn+p2))+log(β2(ωn+p2))].

We can see that the two partition functions are equal! And that is because the ghosts thermalize at nonzero temperature enough to cancel the degrees of freedom obtained from the nonphysical gauge freedom.

Photon Self Energy at Finite Temperature

Let's do some QED. The first thing we haven't addressed so far was the introduction of the chemical potential. As we know from general statistical mechanics, for any conserved charge we can thermally control it in a grand canonical ensemble by introducing an associated chemical potential μ.

Let's do this for the U(1) fermion charge in QED. The conserved current is given for any XX(M) by

J(X)=ψ,Xψ,

where is the Clifford multiplication. In components this is given by

Jμ=ψ,γμψ=ψ¯γμψ.

Therefore the partition function is now

Z=TreβH+μQJ.

This really implies that we have changed the Lagrangian of QED to be (in Feynman Gauge)

Leff(A,ψ,C,C¯)=12FF12δAδA+dC¯dC+ψ,(iDDm+μγ0)ψ.

Here we have incorporated the Gauge fixing term, ghosts, and the fermions ψ with their chemical potential on the charge, and minimal coupling inside DD that is given by eψ¯AAψ.

Once we have the propagators we can actually calculate the polarization by following the standard prescription. Use symmetry arguments (Ward identities, transverse and so on) to constrain the tensorial structure, and then BRST (or what's left or it in the Abelian case) to constrain the rest of the parameters.

So we need the propagators. Luckily we can calculate them with the prescription mentioned above through their regular values. We get

DμνF(k)=gμν(ik2+iϵ+2πδ(k2)f(|k0|))SF(p)=(pp+m)[ip2m2+iϵ2πδ(p2m2)[θ(p0)f+(p0+μ)+θ(p0)f+(p0μ)]].

The interesting thing is that the propagators simply pick up an extra term that corresponds to the distributions f± depending on fermion or boson respectively.

Following the prescription of determining the vacuum polarization we can use symmetry to constrain its tensorial structure like so

Πμν=Πvacμν+ΠTPTμν+ΠLPLμν

where the PT,L are projection matrices to the transverse and longitudinal modes. What that means in practice here is that

P2=PPLPT=0PL(p)p=pPT(p)p=0

so we can express the probabilities for transverse and longitudinal momentum modes. After calculating the actual value and getting the zero modes we finally see that

After calculating these propagators we get the following limits

limp0ΠL(p)=0limp0ΠT(p)=mγ2(β,μ),

where mγ is the effective mass of the photon that at high temperatures has the following limit

limβm0mγ2(β,0)=e2T26.

This result implies screening of the electromagnetic interactions with screening length proportional to β2, which is super interesting.

Spontaneous Illegal Interactions

Having the photon polarization we can calculate s-matrix elements for things that would otherwise vanish. In particular we can have spontaneous pair production! Which I think is pretty awesome.