Académique Documents
Professionnel Documents
Culture Documents
Path Integrals in Quantum Mechanics, Path Integrals in Field Theory, Functional Methods
for Fermions, Path Integrals for Fermions, Path Integrals for Abelian Gauge Theory, Path
QCD Feynman Rules, Color Algebra Techniques, Spinor Algebra Techniques, Some
J. Gunion
Introduction to Path Integrals in Quantum
Mechanics
My notes may differ somewhat in some places from Ryder; they are
based on the notation of the Abers and Lee, Physics Reports.
Introductory Ideas
Let us now divide the time interval between ti and tf into two, with t as
the intermediate time, and q the intermediate point in space.
Repeated application of Eq. (1) gives
Z Z
(qf , tf ) = K(qf , tf ; q, t)dqK(q, t; qi, ti)(qi, ti)dqi (3)
In words, this equation says that the transition from (qi, ti) to (qf , tf )
may be regarded as the result of the transition from (qi, ti) to all available
intermediate points q followed by a transition from (q, t) to (qf , tf ).
or, equivalently,
|iH = eiHt|, tiS . (10)
We also define the vector
0 0 0 iH (t0 t)
hq , t |q, tiH = hq |e |qi. (17)
We will rewrite this propagator using the path integral approach which
will incorporate the quantization of the coordinates.
The first step is to divide up the time interval into n + 1 tiny pieces:
tl = l + t with t0 = (n + 1) + t.
Now notice that for small we can write (dropping the H subscripts to
avoid so much writing):
We now need to substitute many such forms into Eq. (18). This will
yield:
n n +1 n nX
+1
dpk
Z Y Y
hq 0, t0|q, tiH
= lim dqi exp i pj (qj qj1)
n
i=1 k=1
2 j =1
1 o
H pj , (qj + qj1) (tj tj1) , (22)
2
and we have temporarily put back in the hs so that you will recognize
the quantum mechanical action exponential.
Note that the above integration is an integration over the p and q values
at every time . This is what we call a functional integral.
The most important point of the above result is that we have obtained
an expression for a quantum mechanical transition amplitude in terms of
an integral involving only pure c-numbers (no operators).
We can actually perform the above integral for Hamiltonians of the type
H = H(P, Q) (i.e. no products of P and Q). We use square completion
in the exponential for this, defining the integral in the complex p plane
(i.e. by writing p p(1 i) and continuing to the physical situation
by taking 0)
In particular, we have
Z
dp 1 dp 1 1
Z
exp i(pq p2) = exp i(p q)2 + iq 2
2 2 2 2 2
s
1 1
= 1 exp iq 2
2 2
i 2
1
n
Z Y
dqi n nX +1h
1 qj qj1 2
hq 0, t0|q, tiH = lim exp i
2i n
i=1
2i j =1
2
io
qj + qj1
V
2
( Z 0 )
Z t
dq
exp i L(q, q)d , (26)
2i t
where there is actually the extra 21i that gets lost in the nice (but too
simplistic) general notation. In the above,
1
L= q 2 V (q) (27)
2
J. Gunion 230B, 2nd Quarter of Field Theory 13
is the Lagrangian, and
Z
S= L(q, q)d (28)
Let us check that this is the correct answer in the case being discussed.
To do that, we must see how the Schroedinger equation emerges from
our expression. For this, we separate out the very last dqn integral and
write
(q 0 qn )2 q 0 + qn
Z ( !)
0 0 dqn 0
hq , t |q, tiH = exp i iV hqn , t |q, tiH . (29)
2i 2 2
Noting that the terms of order 0 cancel and comparing terms of order
1 we find
2
1
i 0 hq 0, t0|q, tiH = 0 2
+ V (q 0
) hq 0 0
, t |q, tiH (33)
t 2 q
which says that our propagator (which in Eq. (1) we called K(q 0, t0; q, t))
obeys the Schroedinger equation (with m = h = c = 1), as it must
given that it propagates the wave function, see again Eq. (1).
Always keep in mind that the pretty continuous expressions are defined by
the finite interval expressions, and it is only at the end that we can take
0.
0 0 0
hq1 . . . qN , t |q1 . . . qN , ti
N Z t0 XN
dqndpn i
Z Y
= exp pnqn H (p1, . . . , pN ; q1, . . . , qN ) d
n=1
2 h h t n=1
(34)
For the moment, we will continue with some additional results keeping
N = 1.
We will shortly see how it is that time ordering enters into this path
integral game.
Consider first,
0 0
hq , t |Q(t0 )|q, ti
Z Y
0 0
= dqi (ti )hq , t |qn , tn i . . . hqi0 , ti0 |Q(t0 )|qi1 , ti1 i . . . hq1 , t1 |q, ti , (35)
i
where we choose one of the time interval ends to coincide with t0, i.e.
ti0 = t0. If we operate Q(t0) to the left, then it is replaced by its
eigenvalue qi0 = q(t0).
Aside from this one addition, everything else is evaluated just as before
hq 0, t0|Q(t1)Q(t2)|q, ti (37)
in the case where t1 > t2. Well, we have to insert as intermediate states
|qi1, ti1ihqi1, ti1| with ti1 = t1 and |qi2, ti2ihqi2, ti2| with ti2 = t2 and
since we have ordered the times at which we do the insertions we must have
the first insertion to the left of the 2nd insertion when t1 > t2.
Once these insertions are done, we evaluate hqi1, ti1|Q(t1) = hqi1, ti1|q(t1)
and hqi2, ti2|Q(t2) = hqi2, ti2|q(t2) and then proceed as before with
everything else and obtain
hq 0, t0|Q(t1)Q(t2)|q, ti = I (38)
Now, let us ask what the above integral is equal to if t2 > t1? Well, I
assume it is obvious that what you get is I = hq 0, t0|Q(t2)Q(t1)|q, ti.
In short,
I = hq 0, t0|T {Q(t1)Q(t2)}|q, ti . (40)
Clearly, this generalizes to an arbitrary number of Q operators.
L L + hJ (t)q(t) . (41)
Let us define |0, tiJ as the ground state (vacuum) vector (in the moving
frame i.e. with the eiHt included, that is multiplying the Schroedinger
state) in the presence of the source. The required transition amplitude is
where we have omitted a proportionality factor that will not matter in the
end as we will have a ratio of two things with the same proportionality
factor.
Let us define energy eigenstates in the absence of the source term: |ni
with H|ni = En|ni.
Let us define n(q) = hq|ni with 0(q) being the ground state wave
function.
Recall that |qi is the Schroedinger representation state. (We use the
notation |q, ti = eiHt|qi for the Heisenberg state.)
Consider first,
hQ0, T 0|Q, T iJ
Z Z T0
dpdq n o
= exp i [p( )q( ) H(p, q) + J ( )q( )] d
2 T
Z Z
= dq 0 dqhQ0, T 0|q 0, t0ihq 0, t0|q, tiJ hq, t|Q, T i , (43)
Now, we write
hq, t|Q, T i = hq| exp {iH(t T )} |Qi
X
= hq|nihn| exp {iH(t T )} |Qi
n
X
= n(q)n(Q) exp {iEn(t T )} (44)
n
We now substitute these expressions into Eq. (43) and consider the
limit T 0 ei and T ei , where is an arbitrary angle
0 < /2. This is equivalent to rotating the overall time axis so that
it has a downward slant as time runs from to +.
Since the imaginary part of T is i|T | sin and is getting very large, the
term iEnT in the exponential of Eq. (44) is being damped, with the
damping being larger for larger En. The least damped term will be that
with the lowest energy, namely n = 0.
A similar argument implies that in the T 0 case only the m = 0 term
survives in the limit.
h0|0iJ
Z[J ] (46)
i.e. we have an expression for the ground state to ground state transition
amplitude. In getting the above expression, we have used |q, tiH =
eiHt|qiS (but with H and S subscripts implicit) in the following manner:
P
By recognizing that (just insert n |nihn| and take limit)
we can write
J hQ0, T 0|Q, T iJ
Z[J ] = h0|0i = lim . (49)
T 0 ei ,T ei hQ0, T 0|Q, T i
Note that Z[J = 0] = 1! That is, we are always implicitly dealing with
a normalized ratio. In what follows, it is most convenient to simply drop
1 2 1
H H iq , or L L + iq 2 . (51)
2 2
0
implying a damping factor as T enT which isolates the n = 0
state since 0n is smallest for n = 0.
Functional Techniques
This differs from a normal function that maps a number or set of numbers
into a number or set of numbers.
For example, (~ x) maps the set of numbers ~ x into a single number, or
~ x) maps the set of numbers ~
E(~ x into another set of numbers, namely
the electric field 3-vector at point ~
x.
function : Rn Rm , (55)
Then,
F [f ] F [f (x) + (x y )] F [f (x)]
Z
= lim = (x y )dx = 1 . (58)
f (y ) 0
= G(x, z ) . (60)
Back to Z[J ]
We had
Z
Z[J ] = dq 0dq0 (q 0, t0)hq 0, t0|q, tiJ 0(q, t) (61)
inside of which
( Z 0 )
Z t
0 0 J dpdq
hq , t |q, ti = exp i [p( )q ( ) H (p( ), q ( )) + J ( )q ( )]d .
2 t
(62)
Let us now treat, just as we did in the earlier commutator technique for
2nd quantization, the abstract field (x) (for the moment we discuss the
scalar field case) as a coordinate in the sense that we imagine dividing
space up into many little cubes and the average value of the field (x)
in that cube is treated as a coordinate for that little cube (just as we
might use the multi-coordinate generalization discussed earlier).
Then, we go through the multi-coordinate analogue of the procedure we
just considered and take the continuum limit.
The result would be
Z
1
Z
Z[J ] [d] exp i d4x L() + J (x)(x) + i2 (65)
2
hq10 . . . qN
0
, t0|q1 . . . qN , ti
N Z t0 h X
N
dqmdpm i
Z Y n i o
= exp pnqn H({pi}, {qi}) d ,
m=1
2h h t n=1
(66)
where {pi} denotes the set of all the momenta p1, . . . , pN , and similarly
{qi} denotes the set q1, . . . , qN . As usual, such an expression is actually
defined by the discretized version where we go back and discretize the
(67)
L 3 L
p(t) = = 3(t) . (70)
(t) (t)
where
H = L . (72)
We now imagine 2nd quantizing this system by assuming the usual
[q, p ] 3[, ] = i , where and are labels for different
cells of size 3. Making these replacements in our multi-coordinate path
integral generalization that we have shown is equivalent to commutator
quantization yields the form (where {} denotes the set of average
fields values in all the spatial cells labeled by the index )
L
where (~ x, t) = (~
x,t)
the cell average of which is just the (t)
defined earlier.
nZ[J ]
inh0|T {(x1) . . . (xn)} |0i . (75)
J (x1) . . . J (xn) J =0
1
H= ~
2(x) + f [(x), (x)] (76)
2
The result is
Z Z
Z[J ] [d] exp i d4x[L(x) + J (x)(x)] (79)
with
1 1
L(x) = ~
(0)2 f ((x), (x)) + i2(x) , (80)
2 2
J. Gunion 230B, 2nd Quarter of Field Theory 39
where the 12 (0)2 term in L(x) simply comes from the residual
Z
1
exp i d4x 2 (81)
2
obtained above.
As always, the exact form of the proportionality constant will be irrelevant
in our applications.
1
L = L0 + LI , L0 = [ 22] , LI = LI () . (82)
2
1
L0 [ 22 + i2] (83)
2
J. Gunion 230B, 2nd Quarter of Field Theory 40
so that
Z
1
Z
Z0[J ] [d] exp i d4x [ 22 + i2] + J
2
(84)
is the appropriate generating function in the free field theory case.
r
1
1 Y 2i 1 X X
4 4 (K )
Z0[J ] = lim p 8
exp i J 8
J .
0 detK
2
(88)
Here, K 1 is defined by
" #
(K 1
X
4 )
K = , (89)
8 4
1 X Z
4
4
(x y) , 4 d4 x , , (90)
which is equivalent to
the defining equation for the Feynman propagator function. Thus, in the
0 limit we have
1
Z Z
Z0[J ] exp i d4x d4yJ (x)F (x y)J (y) , (95)
2
nZ[J ]
= inh0|T {(x1) . . . (xn)} |0i . (98)
J (x1) . . . J (xn) J =0
Let us check this by showing that h0|T {(x1)(x2)} |0i = iF (x1 x2)
when the l.h.s. is computed using the functional techniques. We have
2
h0|T {(x1 )(x2 )} |0i = i Z0 [J ]
J (x1 ) J (x2 ) J =0
1
Z Z Z
2 4 0 0 0 4 4
= i i d x J (x )F (x x2 ) exp i d x d yJ (x)F (x y )J (y )
J (x1 ) 2 J =0
4 = ~
x x, (102)
41
X X X
4
K + 4J
2
1 T T 1
= [x Kx] + J x (x, Kx) + (J, x)
2 2
1
= (x + K 1J K 1J, K(x + K 1J K 1J )) + (J, x)
2
J. Gunion 230B, 2nd Quarter of Field Theory 46
1 1 1 1
= (x + K J, K(x + K J )) (J, K 1J ) . (103)
2 2
In writing this, we must remember that we are in a real vector space and
that K is a real (in the limit 0) symmetric matrix. As a result, K 1
is also a real symmetric matrix, i.e. K 1 = [K 1]T and, for example,
1 1 T T 1 T T 1 T
(K J, Kx) (K J ) KX = J [K ] KX = J K Kx = J x (J, x) (104)
1 1 1 T T 1 T T 1 1
(K J, KK J) (K J ) J = J [K ] J =J K J (J, K J) . (105)
We shift variables to x = x + K 1J .
The original integral then reduces to
n Z
1 1 1 n i 12 (x,K x)
exp i (J, K J ) 4
d xe (106)
2
n
!
1 (2)n/2
Z X
exp aiyi2 dy1 . . . dyn = Qn 1/2
. (108)
2 i=1 i=1 ai
Now, let A be a diagonal matrix with elements a1, . . . an, and let y be
an n-vector (y1, . . . , yn). Then, the exponent is the inner product:
n
X
aiyi2 = (y, Ay) (109)
i=1
or Z
n/2 1/2
ei(x,Ax)/2(dnx) = (2i) (det A) . (113)
At this point, in our earlier expression we identify A with K, x with x,
and n with the number of cells labeled by to obtain
n Z
11 1 n i 12 (x,K x)
exp i (J, K J ) d xe
2 4
Yr !
1 1 2i 1/2
= exp i (J, K J ) 8
(det K ) . (114)
2
where in the last line I have introduced the normalization factor required
for Z[J = 0] = 1.
Obviously, the expansion in the second line above actually defines the
perturbation theory expansion in powers of LI .
h0|T {(x1)(x2)(x3)(x4)}|0i
h
= iF (x2 x3)iF (x1 x4)
+iF (x2 x1)iF (x3 x4)
+iF (x3 x1)iF (x2 x4) , (123)
2
1 i
Z
exp J (x)F (x y )J (y )dxdy
i J (z ) 2
( Z 2 )
i
Z
0 0 0
= iF (0) + F (z x )J (x )dx exp J (x)F (x y )J (y )dxdy(126)
2
3
1 i
Z
exp J (x)F (x y )J (y )dxdy
i J (z ) 2
i
Z
exp J (x)F (x y )J (y )dxdy (127)
2
4
1 i
Z
exp J (x)F (x y )J (y )dxdy
i J (z ) 2
( Z 2 Z 4 )
2 0 0 0 0 0 0
= 3[F (0)] + 6iF (0) F (z x )J (x )dx + F (z x )J (x )dx
i
Z
exp J (x)F (x y )J (y )dxdy (128)
2
"!# $%&'$($
)+*, -/.1- 032/4657 98:<;>= ,
?A@CBEDGFHBJI ,
K L =NM T Q
O "
P R + T
O + + S98:<;>=
? A C
@ E
B G
D H
F J
B I
+
+
T
+
4
Figure 1: Graphical representation of the interaction; +s indicate J
attachments and closed circle = F (0).
where, in the last line, I have dropped extra O(2) terms from the
denominator expansion.
We must now use this expression for Z[J ] to derive expressions for the
time-ordered products of fields in the presence of the interactions. This
means we must do further functional derivatives.
Let us begin with the 2-point function
2
1 Z[J ]
h0|T {(x1)(x2)}|0i = 2 . (130)
i J (x2)J (x1) J =0
From the above expression for Z[J ], we observe that the [1 . . .] term
simply gives the free particle propagator. The interesting piece is the
Starting from the master form for Z[J ] of Eq. (129), the 1 term in
[1 . . .] simply gives us the free-particle 4-point function that we discussed
earlier.
The next term is
4
i
Z Z
0 0 0 0 0 0
F (0) dx dy dz F (x z )F (y z )J (y )J (x ) exp J F J
J (x1 )J (x2 )J (x3 )J (x4 ) 4 2
J =0
The terms in the last line above are obtained when two of the /J s
act on the external factor (which yields an extra factor of 2 since the
two /J s that act on the external factor can act on either of the
J s appearing in the external factor yielding identical final result), the
3rd /J acts on the exponential, and the final /J acts on the term
brought down from the exponential by the action of the 3rd one above.
The structure of the above is (referring to the 1st term written explicitly)
a free propagator from x3 to x4 times a bubble correction to the
propagator from x1 to x2.
Note that these terms have a disconnected form in which one particle
just passes straight to the final state without interaction.
Diagrammatically, the first term can be represented as
h
i
i x1 z x2 x3 x4 , (136)
2
and there are 5 more terms with the roles of the x1,2,3,4 permuted to
give distinct structures.
The above is the same result as we had in our other approach. It simply
contains the external propagators that connect the external coordinate
W [J ] = i ln Z[J ] (139)
f Lf f Lf
= = a1 + a3C2 , = = a2 a3C1 . (142)
C1 C1 C2 C2
Consistency of the above then requires that the derivative with respect
to one coordinate anti-commutes with the other coordinate since from the
1st and 2nd equations we get
f f
= a3 , = a3 , (143)
C2 C1 C1 C2
f
= (a1 + a3C2) = 0 (145)
C1 C1 C1
by virtue of the fact that the 2nd C1
has no C1 on which to act.
Note that it is also possible to define right differentiation
Rf
= a1 a3C2 , (146)
C1
but we will stick to the L differentiation definition, and will not explicitly
write the L.
implying that
C1 + C1 f = f , or C1 + C1 = 1 (148)
C1 C1 C1 C1
as an operator identity.
Similarly,
f
C2 = a1 C2 , C2 f = a0 C2 + a1 C2 C1 , (C2 f ) = a1 C2 , (149)
C1 C1
implying that
C2 + C2 f = 0 , or C2 + C2 = 0 (150)
C1 C1 C1 C1
as an operator identity.
(Ryder has corrected the small error in the first case present in earlier editions.)
For i 6= j, we will also have {Ci, dCj } = 0, but we must be more careful
for this latter when i = j. Multiple integrals are interpreted as iterated;
for example,
Z Z Z
dC1dC2f (C1, C2) dC1 dC2f (C1, C2) . (153)
where the first equality follows from the standard use of a dummy
integration variable and the other equalitiesR follow
2 from the
R previous
identities. The result is that we must have dC1 = 0 or dC1 = 0.
Obviously, then Z Z Z
dC1dC2 = dC1 dC2 = 0 (155)
since the two integrals are individually 0. As for the non-zero result of
integration, we must be content with a formal definition,
Z
dC1C1 = 1 (156)
Note that Eq. (156) implies that integration and differentiation give the
same result!
d
Z
dC1(p0 + p1C1) = (p0 + p1C1) = p1 (159)
dC1
and Z Z
dC1(p0 + p1C1) = dC1p1bC1 = p1b (160)
!1
1 dC1
Z Z Z
dC1P (C1) = dC1P (C1(C1)) = dC1 P (C1(C1)) ,
b dC1
(161)
i.e. we get the opposite of the commuting result; the Jacobian is the
reverse of what we might have expected.
C
e1 . . . C
en = b1i . . . bni Ci . . . Ci
1 n 1 n
implying that
Z Z
dCen . . . dC ei) = (det b)1 dCn . . . dC1p(C(C))
e1p(C e . (164)
dC
ei
On the other hand, C
ei = bij Cj implies that
dCj
= bij , which is to say
cc + cc = 0 , c2 = 0 , c2 = 0 , (165)
But these rules are not completely consistent with the C1,2 breakup.
Rather, for consistency one should write
Z Z Z Z
dc c = 1 , dc c = 1 , dc = dc = 0 . (168)
However, all that matters in the end are that whatever rules we decide
upon for integration be applied consistently. Following IZ and others, I
will employ the IZ definitions for integration in what follows.
where p1 and p1 are Grassmann and p0 and p12 are regular c-numbers.
Then, Z
dcp(c, c) = p1 p12c = p (170)
1
Z Z
dcdccc exp [cAc] = dcdccc [1 cAc + 0] = 1 + 0 = A.
A
(172)
1
The extra A is what you would have anticipated in analogy with regular
integration by rescaling c Ac and c Ac.
1 2
exp(. . .) [c1A11c1 c1A12c2 c2A21c1 c2A22c2]
2!
1
3 [2c1A11c1c2A22c2 + 2c1A12c2c2A21c1 + . . .] (175)
2!
J. Gunion 230B, 2nd Quarter of Field Theory 74
where the . . . terms are actually 0 by virtue of the fact some ci and/or
ci is squared.
Thus, we have, keeping careful track of signs as we bring the dci or
dci next to its companion,
2
Z
dc1dc1dc2dc2 [c1A11c1c2A22c2 + c1A12c2c2A21c1]
2!
= A11A22 A12A21 = det A . (176)
This generalizes; one always gets det A. For a fairly complete proof, see
Cheng and Li, p. 26 and following.
Another useful result is obtained in our example case by throwing in an
extra factor of, for example, c1c2.
Z
dc1dc1dc2dc2 c1c2 exp [c1A11c1 c1A12c2 c2A21c1 c2A22c2]
Z
= dc1dc1dc2dc2 c1c2 [c1A12c2 + . . .] = A12 = det AA 1
12 . (177)
n
Z Y " #
ckAklcl = det A A
X 1
dckdckcicj exp ij . (178)
k=1 kl
{C(x), C(y)} = 0 ,
C(x)
= 4(x y) ,
C(y)
Z
dC(x) = 0 ,
Z
dC(x)C(y) = 4(x y) , (181)
We will now proceed in analogy with the scalar field case, but using the
Grassmann quantities.
We begin with the standard
and see what happens if we define our generating functional for free
Dirac fields as
Z i
1
Z
4
h
Z0 (, ) = [d ][d ] exp i d x (x)(i m) (x) + (x) (x) + (x) (x) , (183)
N
Let us write
Z Z
d4x(x)(i m)(x) = d4xd4y(x)O(x, y)(y) (184)
with
O(x, y) = (i x m) 4(x y) (185)
which results in the structure for the stuff under the integral in the
exponential above of the form (not writing all the integrals in x, y and
)
+ 1 1 + 2 2 + 1 1 + 2 2
"
1 2
Z
= d 1 d1 d 2 d2 i 2 1 O11 1 2 O22 2 + 1 O12 2 2 O21 1
2!
1 3
+ i 6 1 O11 1 2 2 2 2 + 2 O21 1 2 2 1 1 + 1 O12 2 1 1 2 2 + 2 O22 2 1 1 1 1
3!
#
1 4
+ i 24 1 1 2 2 1 1 2 2 + terms that give 0
4!
1 1 1 1
= det(iO ) exp i 2 O22 2 i 2 O21 1 i 1 O12 2 i 1 O11 1
X 1
= det(iO ) exp i k O (188)
kl l
kl
So, the
above result
is a purely algebraic result. Now we wish to compute
h0|T (w), (z) |0i.
We then have
1 1
h0|T {(w), (z)}|0i = Z0
i (w) i (z) = =0
1
Z
4
= () d x(x)SF (x z)Z0
i (w) = =0
= iSF (w z) (194)
Overall, you could say that we have replaced one arbitrary (except that
causality, . . . seems to require it) assumption of the {(x), (y)} anti-
commutation condition with another assumption namely, Grassmann
variable integration for fermionic degrees of freedom.
Both are equivalent, as we have seen in the simple case above, and will,
among other things, lead to a causal theory.
One final note: we will employ this same formalism eventually to rewrite
a determinant, that will arise during gauge theory 2nd quantization (as
part of the gauge fixing process), in terms of an integral over (fictitious,
but very useful) spinless, anticommuting scalar Grassmann objects c and
c.
Let us see how the need for the gauge fixing condition arises in the path
integral context.
Naively, we might start with L = 14 F F where F = A A
as usual, and employ
Z Z
[dA] exp i d4x(L(A) + JA) (196)
Lets examine what happens more closely. We rewrite, using the usual
partial integration technique and the antisymmetry of F ,
1
Z Z
4 4 2
d xL(A) = + d xA(x) g A (x) . (197)
2
Let us define
K = g 2 . (198)
R
As in the scalar field case, we would like to perform the [dA] by
completing the square using (K )1.
The problem is that (K )1 does not exist. We have already discussed
this, but lets just quickly review.
There are 4 different ways of checking that the inverse does not exist.
1. First, we can show that K K = 2K 2 which means that K is
a projection operator. A projection operator has no inverse.
2 1
(g k kk )K (k) = g . (200)
1
The most general form of K (k) is
1
K (k) = a(k2)k k + b(k2)g . (201)
det(KK) = det K,
(det K)2 = det K (203)
which has only the two solutions det K = 0 or det K = 1. The latter
holds only for the trivial K = 1 choice.
4. Finally, that det K = 0 can be also seen by noting that K has a zero
eigenvalue for the trivial (gauge transform) function (x):
(g 2 ) = 2 2 = 0 . (204)
The simplest way to get around this is to impose the gauge condition by
hand in L. If, for example, we impose A = 0 directly we have
4 1
Z Z
4
d xL = d x A( 2g )A
2
1
Z
d4x A 2g A . (205)
2
1
L= 41 F F ( A)2 (208)
2
1
Z Z
d4x 14 F F = A( 2g )A ,
and
2
1 1
Z Z
d4x ( A)( A ) = d4xA A (209)
2 2
so that the gauge condition term cancels the offending part of the original
stuff, leaving us with the simple form employed in the previous discussion.
1
L = 14 F F ( A)2
2
1 1
00
=00 A 2g + 1 A , (210)
2
00 00
R
where = means under d4xL.
Then, the kernel in question is (in Fourier transform space)
1
K (k) = k2g 1 k k (211)
1
whose inverse (defined by K(k)K (k) = g ) is
1 1 k k
K (k) D (k) = 2 g + ( 1) 2 . (212)
k k
Just to review the square completion process, let us refer to our starting
form Eq. (196), with L being the L after including gauge fixing. We
4 1
Z Z
4
d xL(A(x)) = d xd y A (x)K (x, y)A (y)
4
(215)
2
and
1
(K 1J )nKnm(K 1J )m = Knl 1
JlKnmAmKmp Jp
1 1
= JlKln KnmKmp Jp by symmetry of K 1
1
= JllmKmp Jp
1
= JmKmp Jp . (218)
( )
1
Z Y X X
[dA0p] exp i 4
4
A0nKnmA0m
p n m
2
r
1 Y 2i
= , (219)
det Knm p
8
1
X
4
Kpm nm
Knp = (221)
p
8 4
K(x z) = x2 4(x z)g (224)
x2 g D (x y) = 4(x y)g (225)
d4 k g
Z
ik(xy )
D (x y) = 4
e , (226)
(2) k2
where D (xy) is the result that we obtained for the gauge propagator
(see above example) using the commutator 2nd quantization techniques.
Note that we have obtained the correct sign and there is no factor of
1
2
because each of the J
derivatives can see either of the J s in the
exponential.
where we have used the fact that S(A) = S(A) on a given gauge
orbit. The latter factor is the infinity that gives us the problem. We
modify it by instead writing
Z Z
i
(f (A ))2
[d] [d]e 2 . (230)
By integrating over all possible values for the given function f at each
space-time point, we have obviously eliminated any sensitivity to the
value of f for a given field configuration A.
You may ask what kind of determinant it is that appears above? Well, to
understand you must recall that these functional integrations are really
defined by dividing space-time up into cells. Since f typically contains
a derivative, it is actually sensitive to differences between neighboring
cells. Thus, what we really have is f i
j
where the i, j indicate the cells
considered. This creates a matrix that we shall call Mij that records how
fi in cell i responds to a gauge transform in cell j. We will return to this
M shortly.
For the moment, let us note that if we do the above, we obtain
Z
1 f
Z Z
Z[J ] [dA] [d] exp i d4x L + J A f 2 det ,
2
(232)
f (x) 1 1
M (x, y) = = (x2 (x)) = x2 4(x y)(233)
.
(y) e (y) e
Let us recall that for every conserved quantum number one can construct
a transformation on the fields which leaves L invariant. The simple case
we have dealt with so far is charge. Let us review.
If we say that i has charge qi, this means that under the U (1) group
1 . . . n ei(q1+...+qn) 1 . . . n . (238)
P
If L is invariant under the transformation, then i qi = 0 is required,
which is to say that the product is neutral.
If does not depend upon x, then each of the kinetic energy terms
i i is also invariant.
The corresponding infinitesimal transformation with = 1 is
i = iqii (239)
i.e.
X L
J = 0, with J = i qj j . (241)
j
(j )
R
Further, Q = d3xJ 0 is a conserved charge operator, which generates
the transformation: i.e.
i = i[Q, i] (242)
L
= 0j (243)
(0j )
implying
XZ
Q= d3x i0j (x)qj j (x) . (244)
j
Then,
XZ
[Q, i(y)] = d3xqj j (x)[0j (x), i(y)]
j
XZ
= d3xiqj j (x)[iij 3(x y)]
j
= qii(y) , (245)
implying
i(y) = i[Q, i(y)] = iqii(y) (246)
i
= ~ ~
doublet , i = ijkj k triplet . (251)
2
f 123 = 1 ,
a
Ta = (254)
2
Di = ( + ieqiA)i (257)
LA = 41 F F , F = A A (261)
b a b a
eiL eiL eiL eiL = 1 + 2[La, Lb] + . . . (262)
The cabc are called the structure constants of the group. You might
worry that [La, Lb] = icabcLc would not be sufficient to guarantee the
matching to still higher orders in . But, in fact, it is all you need.
By the definition and the expansion process above, the cabc are antisymmetric
in the first two indices. Further, they are real if the La are hermitian.
Reality of the cabc follows from [La, Lb] = icabcLc and the hermiticity
of the Las by taking the hermitian conjugate of this relation:
As you see, the cabc are determined by the group multiplication law.
They also determine the multiplication law as follows:
ia La i b Lb i c Lc
e e e (265)
Representations
So, we have learned that the generators and their commutation relations
define the Lie algebra associated with the Lie group.
[T a, T b] = icabcT c , (270)
D is a real (since the Las are hermitian) symmetric (from the cyclic
property of the trace) matrix, so we can diagonalize it by choosing
appropriate real linear combinations of the Las.
In more detail, and using subscripts to simplify my typing, let us write
L0a = MabLb. Then
In addition, since the La are hermitian (for the compact Lie groups of
interest to us), all the kas are positive.1 This is most simply seen
by noting that any hermitian matrix can be diagonalized and that the
diagonal will then contain real eigenvalues, i (i = 1, . . . , n). In the
diagonal form for a given La, not all the i can be zero (otherwise it
would be equivalent to the 0 matrix) and we thus have
X
a a a a
L L = diag(21, . . . , 2n) Tr(L L ) = 2i = ka > 0 . (275)
i=1,...,n
Conjugate representation
(1 i aLa
r ) (278)
(1 + i a(La
r ) )
(279)
[T i, T j ] = icijkT k . (280)
As before,
~ ~ ~
(x) 0(x) = eiL (x) (1 iL
~ (x))(x) , (281)
~ A
D = ig L ~ (x) (282)
0 0
D (x) = (1 iL ~
~ (x))D (x) . (283)
0 0 j ~
D = ( igA0 Lj )(1 iL
~ )
~ iL
~ )
= (1 iL ~ ~ igA0 j Lj (1 iL ~
~ )
= (1 iL ~
~ )( j j
igA L ) (284)
which, after canceling common terms on the two sides of the equation
becomes (since it must hold for arbitrary )
iL ~ igA0 j Lj (1 iL
~ ~ = igAj (1 iL
~ ) ~ j.
~ )L (285)
j 1
A0 = Aj j + cjkl kAl . (286)
g
j j 1 ?
iL ig(A j +cjkl kAl)Lj (1iLn n)=igAj(1iLn n)Lj
j
.
g
(287)
Neglecting terms of order 2 in the infinitesimal limit, this reduces to
j j j n n j j j n n j j jkl k l j
iL igA (1 iL )L igA [L , iL ] + i L igc A L
? j n n j
= igA (1 iL )L (288)
where our main manipulation was to commute two Ls past one another
on the left hand side. Canceling common terms, the above reduces to
?
igAj[Lj , iLn n] igcjnl nAlLj =0 . (289)
But, we know that [Lj , Ln] = icjnlLl and the above reduces to
?
igAj(i nicjnlLl) igcjnl nAlLj =0 , (290)
D D (1 iL ~ D ,
~ )D (296)
and
DD (1 iL ~ D .
~ )D (297)
Then, if we define
(DD D D) igLj F
j
(298)
0 j ~ = (1 iL
~ ) ~ jF j
~ )L
Lj F (1 iL (300)
which in turn implies that (assuming above must hold for arbitrary )
0 j ~ j F j (1 + iL
~ )L ~ = U ()L
~ ) ~ U 1() (301)
~ F
Lj F = (1 iL
~ F
(dropping terms of order 2 and defining L ~ Lj F j ). Using this,
we can show that
h i 1 j j
~ F
Tr (L ~ )(L
~ F ~ ) = F F (302)
2
is invariant under gauge transformations.
Proof:
1 0 j 0 j
h i
F F ~
= Tr (L ~ 0 )(L
F ~ ~ 0
F )
2
J. Gunion 230B, 2nd Quarter of Field Theory 131
h i
~ F
= Tr U (L ~ )U 1U (L ~ )U 1
~ F
h i
~ F
= Tr (L ~ )(L
~ F~ )
1 j
= F F j (303)
2
Returning to
0 j ~ j F j (1 + iL
~ )L ~
~ )
Lj F = (1 iL (304)
we see that
j
F Lj = i m[Lm, F
k
Lk] = i mF
k
icmkj Lj (305)
or
j
F = cjmk mF
k
. (306)
A vector that transforms in this way (i.e. using the group structure
constants) is transforming like a member of an adjoint representation of
j m m k
F = iTjk F = i(icmjk) mF
k
= cmjk mF
k
(307)
( ig L ~ 0 )U = U ( ig L
~ A ~ A
~ ) , (309)
(U ) + U ig L ~ 0 U = U igU L
~ A ~ A
~ (310)
i
L A = U L AU (U )U 1 .
~ ~ 0 ~ ~ 1
(311)
g
0
igF 0 = (D
0
D0 D0 D
0
)0 = U (DD D D) (313)
0
igF U = U (igF ) (314)
0
F = U F U 1 . (315)
h i 1
Tr [F F ~ F
] = Tr (L ~ )(L
~ F ~ ) = j
F F j (316)
2
1
m2AjAj = Tr[m2(L
~ A
~ )(L
~ A
~ )] (317)
2
J. Gunion 230B, 2nd Quarter of Field Theory 135
would not be invariant under the transform of L~ A
~ given in Eq. (311)
~ A
because of the extra term that is not simply U L ~ U 1 .
1
|Rea0| < (318)
2
u de+e . (319)
You will learn, or perhaps have already learned, that at low energies
such interactions are well-described by the Fermi-theory in which the
Lagrangian for the above interaction is written as
GF
Lweak = dPLu PLe , (320)
2
1 1
LA = 14 F
j
F j ~ ~ ~ ~
= Tr[L F L F ] = Tr[F F ]
2 2
(322)
~ F
where F = L ~ = Lj F j was defined by
i i
F [D, D ] = [ igA, igA ] , (323)
g g
which yields
F = A A ig[A, A ] (324)
~ A
where A = L ~.
LiF
i
= Li(Ai Ai) ig[Lj Aj, LkAk ]
= Li Ai Ai + gcjkiAjAk , (325)
which implies [using cjki = cikj , as derived earlier, see Eq. (294)]
i
F = Ai Ai gcikj AjAk . (326)
Equations of motion
The equations of motion are easily obtained from LA above. It is easiest
to start from first principles and consider
Z
0 = d4xTr[F F ]
Z
= 2 d4xTr [F ( A A ig[A, A ] ig[A, A ])]
Z
= 4 d4xTr [F ( A igAA + igA A)] . (328)
Tr[F AA ] = Tr[AA F ]
F ig[A , F ] = [D , F] = 0 . (331)
[D 0, F
0
] = U [D , F]U 1 = 0 . (332)
~ A
igF = (ig L ~ ) (ig L
~ A
~ ) g 2 [L
~ A
~ , L
~ A
~ ] (334)
which we rewrite as
~ A
F = (L ~ ) (L
~ A
~ ) ig[L
~ A
~ , L
~ A
~] . (335)
i 1 i 1
F = [ ( U )U ] [ ( U )U ]
g g
i 2 h
1 1 1 1
i
ig ( U ) U ( U )U ( U ) U ( U )U
g
i h 1 1 1 1
i
= ( U )( U ) ( U )( U ) ( U )( U ) + ( U )( U ) = 0 .(336)
g
In order to get the form after the 2nd =, we used two identities for the
last two terms of the first form. For example, from
0 = (1) = (U U 1) = ( U )U 1 + U ( U 1) , (337)
implying
( U )U 1 = U ( U 1) (338)
which in turn implies
(U )U 1( U )U 1 = (U )U 1[U ( U 1)] = (U )( U 1) .
(339)
Proof: Direction 2
where y > z means y is further along the curve C 0 than z. Well, there
is still work to do to show that the analogue of Stokes theorem gives
Z Z
dy A(y) = dy dy (y A(y)
y
A(y)) , (344)
C0 S0
dropping terms that are of order (area)2. In short, the quadratic term
reduces to
Z
dy dy (A(y)A(y) A(y)A(y)) (350)
S0
As a result, we find
Z
U 0 1 + ig dy dy F(y) , (351)
S0
where
F = A A ig[A, A] (352)
is the usual definition of F as given in Eq. (335). Since we have assumed
F = 0 (throughout all of space), U 0 1.
The difference between two longer paths can be realized by the difference
between many smaller paths, or equivalently the surface filling in the
i
A(x) = (U )U 1 (353)
g
I will employ a treatment that is a bit different than found in Ryder, Sec.
3.6. Presumably by looking at both you will learn the most.
In a curved space, to compare a vector field at two different space time
points, V(x0) vs. V(x), you must first parallel transport V(x) from
x to x0, and then reference both of them to a given coordinate system
defined at x0. The result will take the form
DV = dV V . (355)
/
This is illustrated in Fig. 2.
"!#%$&'
0 2143 5146 0 *89:;3
6
3
89 0 2143
E
6 <0
/>= 2143 @? A3
BDCFE ;G5,H E *
H
V =
V
dx
, V = +V dx , (356)
we have
DV = (V +
V
)dx
(358)
The signs and conventions of Eq. (356) are chosen so that
0 = (VV ) = VV + VV
= V(
V dx ) + ( V dx )V
= VV dx(
+
) (359)
VP P1P2 = ( V )P b + (V )P1 (a + a)
VP P3P2 = (V )P a + ( V )P3 (b + b ) . (361)
(V )P1 = ( + b )P (V +
V b
)P
( V )P3 = ( + a)P (V +
V a
)P . (362)
From now on, the P subscript is omitted for quantities evaluated at point
P . We also use
(a + a) = a
a b
b + b = b
b
a . (363)
We obtain
V = ab V +
where the contributions from the a and b terms in Eq. (363) ended
up canceling by virtue of the symmetry
=
.
In general relativity R would be called the Riemann-Christoffel
curvature tensor. The idea is that if the transport around the closed path
ends up producing a change in V then the space-time truly is curved
and the non-zero value of R signals the presence of true curvature.
In particular, it is possible for the connection to have a non-zero
value even if there is no true curvature.
DV = (V +
V )dx
(365)
~ j A
Dj = j ig(L) ~ k . (366)
k
jk = ig(L)
~ j A
k
~ = ig(A)j
k (367)
in the internal charge space, i.e. nothing to do with space-time for the first
two indices. Thus, it is as if the gauge fields establish a curved system
of internal coordinates that must be used in computing how an internal
space vector changes as one moves in real space-time. The gauge field
provides the instruction as to how to compensate for the change of the
local internal index frame in going between different space-time points
and
j = jkkdx = ig(L) ~ j A
k
~ dx (368)
under parallel transport. With this identification, the curvature tensor
becomes
j j j j j
Rk = k k + lk l lkl
j (x) j (x + dx)
= ~ j A
j (x) + j (x) = j (x) + ig(L) ~ dxk(x)
k
For a finite interval, the expression for the parallel transport matrix P
becomes ( Z 0 )
x
P (x0, x) = P exp ig ~ A
L ~ (y)dy . (372)
x
~ A
or, using the standard notation A = L ~ ,
(which follows simply from the definition of the commutator and leads to
the structure constant relation cijmcmkl + cjkmcmil + ckimcmjl = 0) is
equivalent to
DR + DR + D R =0 (380)
Consider Z Z
Z= dxdyeiS (x,y) = r eiS (~r) .
d~ (381)
Suppose S(~
r ) = S(~
r), where ~
r = (r, ) and ~
r = (r, + ).
For example, S = S(r).
R
In this case, it is clear that there is a factoring volume factor d = 2.
We will divide out this factor in a fancy way that will be closely analogous
to what we do in the gauge theory.
R
1. Write 1 = d( ).
2. Insert this into Z:
Z Z Z
Z= d d~ r eiS (~r)( ) = dZ , (382)
5. is invariant since
Z
r0 )]1 =
[g (~ d[g(~
r+0 )]
Z
= d00[g(~
r00 )]
r )]1
= [g (~ (387)
The final very crucial point is that Z is not only independent of , but
it is also independent of the form of g (so long as g(~ r) = 0 defines a
unique point on each orbit).
The proof begins with the definition
1
g(~
r)
Z
1 = g (~
r) d[g(~
r)] = g (~
r) (392)
g =0
R
To proceed, we use h(i) d~
r =i r drd,h (ii) S(~ ri) = S(r), (iii) d =
d( + ) and (iv) g(r, +) g (r, +)
R
= ( +)
all that matters
g =0 g =0
g(~
r)
Z
Z = r drdeiS (~r) [g(~
r)]
g =0
g(r, + )
Z
= r drd( + )eiS (r) [g(r, + )]
( + ) g =0
Z
= r dreiS (r) (394)
where the last equality simply follows from the chain rule
g(r, + )
Z Z
d( + ) [g(r, + )] = dg[g] = 1 . (395)
( + ) g =0
~ (x)) = 0 ,
fa(A a = 1, 2, 3 for SU (2) group . (396)
~ ~ (x)) = 0 ,
fa(A a = 1, 2, 3 (397)
U () = 1 iL ~ + O( 2)
~ (398)
and defining
3
Y
d = da . (399)
a=1
This exhibits an important invariance, namely
d( 0) = d 0 (400)
where the left-hand side is the notation for the delta functional.
The implication of the above formula can perhaps better be appreciated
by writing it in the form
~ ] = det Mf
f [A (405)
where
~ ~
fa(A(x))
(Mf )ab(x, y) = . (406)
b(y)
fa =0, a=1,2,...
The determinant is thus a determinant over all a, b and over all (x, y)
clearly something that can only be defined by dividing space-time up
into many cells in the usual fashion.
Another phraseology is that Mf is the response of the fa[A~ ] to the
~ ] = 0.
infinitesimal gauge transformation away from fa[A
where the first function is for fa=1 and the 2nd is for fa=2. We now
perform the 1 integral using the first function to get
12 2
M
1
Z
2 21 22 2
= 11
d M 11
+ M
M M
Z
2 22 11 21 12 2
= d [M M M M ]
1
= . (408)
det M
This same simplified example applies equally well to a one-parameter
1
Aa
= Aa
+ abc b
Ac a (409)
g
Let us see how this works in a particularly simple case: The Axial Gauge.
1. We choose fa = Aa
3 = 0 (a = 1, 2, 3 for SU (2)) to define the gauge.
1
~ ~ (x)) = Aa(x) + abc b(x)Ac(x) 3 a(x)
fa(A 3 3
g
1
= 3 a(x) since A ~3 = 0 . (411)
g
1
[Mf (x, y)]ab = 3x 4(x y)ab . (412)
g
4. The axial gauge picks out a unique element on each gauge trajectory
provided the fields vanish fast enough at .
The logic is as follows.
(a) Gauge transforms, U , which preserve A ~ 3 = 0 are independent of x3.
(b) If the fields vanish at then U must keep the fields vanishing at
, implying that U 1 at in all directions.
(c) If U 1 at and U is independent of x3 everywhere, then U = 1
everywhere.
The axial A ~ 3 = 0 gauge is, in fact, a very useful gauge not only
because det Mf 6= 0 but most importantly because Mf is independent
of the gauge field. This makes it a very simple gauge to use in many
applications.
In fact, in this gauge it is possible to show that canonical quantization
and path-integral quantization are completely equivalent.
Then, since physical results are independent of gauge, we conclude
that any canonical quantization will give the same physical results as
any path-integral quantization (with gauge fixing incorporated in both
techniques).
We probably do not have time to go through this proof and so I will skip
it for now. It can be found in Cheng and Li, with a related proof in a
different gauge in Abers and Lee.
and
Z
~ ~ ]1 = 0 ~ (~0 )
f [ A [d (x)][fa(A (x))]
Z
= ~ (~0)(x))]
d[ 0(x)(x)][fa(A
Z
= ~ (~00)(x))]
d[ 00(x)][fa(A
~ ]1 ,
= f [A (414)
(415)
where in the one step weR used the gauge invariance of f , of L and of
the integration measure [dA ~ ] 2 and in the next step we simply shifted
~ ~ A
to A ~ 0 and then dropped the prime.
~
R
We can now factor out the infinite quantity [d(x)], since everything
~ and define
else has no dependence on ,
Z Z
Zf [J ] = ~ (x)]f [A
[dA ~ ][fa(A
~ )] exp i d4x[L(A(x)) + J~ A
~ ]
(416)
and it is always useful to keep in mind the equivalences (ignoring indices):
f
~
f [A] = det Mf = det . (417)
f =0
2
If we are integrating over all vector field values, that is equivalent to integrating over all vector field values
transforming by some given . Either way, the coverage of vector field values is complete.
As stated, we do not have time for this proof. I must ask you to assume
that such a proof exists. You can attempt to penetrate the version given
in Cheng and Li and the related one in Abers and Lee. (Proof in my
notes is found in 245B gray notebook, a summary of which appears in an
Appendix which follows this section.)
Assuming that the equivalence between the 2nd quantization and path-
integral approaches can be established in axial gauge, the only remaining
3 ~ ] since it is field-independent. As we shall see, in general we cannot
In the axial gauge, we can drop f [A
~ . The ability to drop f [A
drop this factor since it will depend on A ~ ] is the reason why the equivalence between
2nd-quantization and the Faddeev-Popov path-integral procedure is relatively straightforward to establish in axial
gauge.
is independent of f .
Proof:
1. Write [dA] as follows.
Consider the A which are equivalent by gauge transform (i.e. those
on the same gauge orbit).
Choose a A b on each orbit.
Then, all of the A are generated from the collection of A ~
b by U ()s.
This means that
[dA] = [dA ~
b][d] (420)
which is a finite integral in which we are able to employ the same sample
Ab from each gauge orbit regardless of what choice we make for the gauge
functions fa(A).
I hope that you recall the simple analogue result that we had in the
2-d circular orbit picture. There we showed that the gauge-fixed path
integral specified by a function g was actually independent of the choice
of g (so long as g specified a unique location on each of the circular
orbits) and that for any g one obtained r dreiS (r) when S(~
R
r ) = S(r)
was independent of location on the orbit.
Y T DY
Z
= dy1 . . . dyN e , with Y = RX (The Jacobian is 1.)
These formulae are valid when the determinant does not vanish. If it
does, it means that one or more of the di are equal to zero, leading to
infinities from integrating over the associated infinite intervals. But, we
can get a sensible answer even if the determinant vanishes. We simply
have to remove the culprit infinite integral(s). Assume there are n zero
eigenvalues. We define the restricted Gaussian integral
Z
X T (y )AX (y )
Grest(A) = dy1 . . . dyN ne , (429)
(430)
to obtain
N
! N
y
Z
X T (y )AX (y )
Y Y
Grest(A) = dxi det (ya(x))e .
i=1
x a=N n+1
(432)
This integral is perfectly well-defined.
Q The ya are arbitrary functions of
y
the xi, and the extra factors det x (y) in the measure effectively
restrict the integration from an N -dimensional space to an N n
dimensional one. As the construction has shown, Grest(A) does not
depend on the specific form of the ya(x) so long as they are sufficiently
cleverly chosen so as to do the job, i.e., restrict the integration region; if
y
they do not, the Jacobian det x will be zero.
We have seen that the path integral generating function in axial gauge
can be written as:
Z
Z[J~] = ~ ](A
[dA ~ 3) exp{iS(J~)}
Z
= ~ , J~)}
[dA0][dA1][dA2] exp{iS(A (433)
with Z
~ , J~) = 1
S(A a
d4xF F a + J~ A
~ . (434)
4
We will shorthand and write S(J~) for the above in what follows.
where
1 a
Z
0 4 1 a a a a abc b c a a
S (J~) = d x 4 F F F A A + gc AA + J A
2
(436)
a
If we integrate over F using square completion we see that Z 0[J~] is
equivalent to Z[J~] (dropping irrelevant constant multiplying factors as
usual).
0 1 F a F a 1 F ij a Aa Aa + gcabc Ab Ac
L = 4 i j j i i j
2
1 0i a a a abc b c
i3 a a 03 a a
F 0 Ai i A0 + gc A0 Ai F (3 Ai ) + F (3 A0 )
2
(437)
where i, j = 1, 2!
L0 L0 L0 L0
= , = (438)
( F
a ) a
F ( Aa
) Aa
give rise to the following constraint equations (i.e. ones with no time
derivatives):
a
Fij = iAa
j j A a
i + gc abc b c
Ai Aj (439)
Fia3 = 3Aa
i (440)
F0ai = 0Aa
i i A a
0 + gc abc b c
A0 Ai (443)
Fi
a
= gcabc Fijb
Aj c + F0biA0 c . (444)
as we shall verify later. For now, we wish to first show that this result is
equivalent to Z 0[J~] obtained earlier from the path integral point of view
and defined in Eq. (435)
So, we now use this Lemma to rewrite our path integral form by using
it for the [dFij ], [dF03], [dFi3] and [dA0] integrations appearing in Z 0[J~]
defined earlier in Eq. (435). The F integrals are of exactly the form
0i a 1
0Aa iAa abc
Ab0 Aci (3Aa 2
L3F i 0 + gc 0 ) (447)
2
where it is the last term that comes from the F 03 integral (equivalently
a
Lemma substitution of F03 = 3Aa 0 ). So now we must do (after doing
the previously mentioned parts integrations to get iF 0i a and also doing
a parts integration to get 3Aa a a 2 a
0 3 A0 A0 3 A0 ) the following path
This is exactly the form you would have gotten by eliminating A0 in the
starting form Eq. (447) using the constraint equation, Eq. (442),
iF0ai 3(F03
a
) = iF0ai 3(3Aa
0 ) = gc abc b
F0i A ic
. (449)
To see this, rewrite Eq. (447) in the form one obtains after the above
partial integrations (and a switch of the dummy summation indices
0i a 0i b bac 1
L 3 (iF + gF c Aci)Aa
0 + Aa 2 ab b
A0
2 0 3
a 1
J Aa
0 + Aa 2 a
0 3 A0
(same J a as defined previously) (450)
2
32Aa i
F0ai bac
F0biAi c = J a
0 = + gc (451)
to obtain
1 1
Z Z Z
1
d4xL 3 d4x J a(x)Aa 0 (x) = d 4
xd 4
y J a
(x)Oab (x y)J b
(y)
2 2
(452)
1
where we used Aa b
R 4
0 (x) = d yOab (x y)J (y) as follows from the
constraint equation written in the form Eq. (451) when solved for A0 in
integral equation form. The above result is that obtained by path integral
techniques as displayed in Eq. (448).
The final element of the proof is to work more on the 2nd quantization
form. In particular, recall that the Hamiltonian formulation of 2nd
quantization is equivalent to the Hamiltonian Path Integral formalism
when we have independent fields and their canonically conjugate momenta.
In the simplest scalar case this was the statement that using
Z Z
Z[J ] [dd] exp i d4x[ H(, ) + J ] (453)
to develop Wicks theorem results was the same as using 2nd quantization
to obtain the Wicks theorem contractions.
(460)
Z
Z 0[J ] = ~ 3) exp{iS 0(J )} .
[dF ][dA](A (461)
R (462)
5. But, we were able to show that performing the [dA0] converted
S 0(J, with F s substituted) into S 00(J ). Equivalently, we showed that
substituting for the Aa 0 using the A a
0 constraint equations, but very
carefully treated (we had to use O 1 to complete the substitution),
gave us S 00(J ).
Possibly useful references beyond Ryder for the material presented here
are:
1. Cheng and Li p. 248 and following;
where you will notice that I did not change f [A~ ]. That is, the above
change assumes that the response function Mf is independent of this
shift of the gauge condition. In fact, what is important is that the form
of f [A~ ] as a function of A
~ is independent of the shift of the gauge
condition.
For the particular change of the gauge-fixing condition we are discussing
here this is indeed the case since Ca(x) is independent of A ~ this
~ dependence of the operator
shift generalization does not affect the A
Mf , which is the local response of the gauge condition to a small gauge
yielding
( Z " #)
1
Z Y
d 4 a a 2 ~
X X
Z [J ] = ~ ] exp i
[dA ] det Mf [A d x L(A (x)) + J A fa (A) .
a 2 a
d
(467)
~ dependence of det Mf .
where I have made explicit the (possible) A
The next manipulation is to use our trick for writing det Mf , see
Eventually we will want to introduce some source currents for the ghost
fields ca and ca, and use generating function techniques for them. But,
The first very important choice is the covariant (Lorentz) gauge which is
defined by taking
~ ) = Aa ,
fa(A a = 1, . . . , N . (471)
For small as, we have (using the notation f abc for the group structure
constants)
1
Aa
(x) = Aa
(x) +f abc b
(x)Ac(x) a(x) (472)
g
so that
1
Aa
= Aa
+
f abc b
(x)A c
(x) a
. (473)
g
1 nh i o
Mf (x, y)ab = ab gf abcAc 4(x y) . (475)
g
Before proceeding, let us note that this response function has a form (as
~ ) that is in fact independent of the value of Aa from
a function of A
which one starts as promised by our general proofs.
Z Y Z
Z[J, , ] = [dAd][dcd][dcd] exp i d4xL(x) (477)
d
with
1 2
L(x) = 14 F a F
a
Aa
2
nh i o
ca ab gf abcAc cb
where
a
F = Aa A a
+ gf abc b c
A A . (479)
In the above, everything is a function of x and repeated indices a, b, c
and , in the L expression are summed over.
Of course, this is for the pure gauge theory only. Fermions and other
particles are brought in as needed.
In the above, the ghost fields are fully parallel in function to the gauge
fields. The Feynman rules must include propagators and vertices involving
the ghost fields as well as the gauge fields.
We will derive the relevant Feynman rules using the functional techniques,
but at this point they can equally well be read off of the Lagrangian
using naive procedures where we imagine all the gauge and ghost fields
to be 2nd quantized operators. In other words, we can use the path-
integral technique to develop the required additional ghost fields and the
associated Lagrangian form and then treat all fields as 2nd quantized
operator fields.
We begin by separating the full L given above into the free-particle and
the interaction components.
We then compute the Z0[J, , ] for the free-particle components and
then bring in the interactions via the perturbative expansion.
We have
Z0 [J, , ] = Z0 [J ]Z0 [, ]
Z
1
Z Y
b 4 a a 2 a 2 a a
= [dA ] exp i d x 41 ( A A ) ( A ) + J A
2
b
Z Y Z i
a a
h
b b 4 a 2 a a a
[dc ][dc ] exp i d x c c c c . (480)
b
To compute Z0, we can separately perform the square completion process for
the Ab fields and the cb, cb fields.
For the former, we just have the addition of a group index to carry around
with our cell and Lorentz indices. The full collection is , , a, where
is the cell label.
However, you can see that the quadratic part of L is completely diagonal
in the group index a. Thus, it should not surprise you that we get the
where
2 1 4
Kab (x y ) = g x 1 x x (x y )ab (484)
has the inverse Dab defined by
Z
4 bc c 4
d zKab (x z )D (z y ) = g a (x y ) (485)
with solution
d4 k ik(xy ) g k k
Z
Dab (x y ) = ab e 2 + (1 ) . (486)
(2 )4 k k4
d4 k ik(zy ) g k k
1
Z Z
4 2 4
= d z g x 1 x x (x z )ab bc e + (1 )
(2 )4 k2 k4
d4 k g k k
1
Z
2 ik(xy )
= ac g x 1 x x e + (1 )
(2 )4 k2 k4
d4 k ik(xy ) g k k
1
Z
2
= ac e k g + 1 k k + (1 )
(2 )4 k2 k4
d4 k ik(xy ) k k k
" #
k k 1 k 1
Z
= ac e g (1 ) 1 + (1 ) 1
(2 )4 k2 k2 k2
d4 k ik(xy ) h
Z i
= ac e g 0
(2 )4
4
= ac g (x y ) , (487)
as hoped.
Going to momentum space, the above form of Z0[J ] generates the
Feynman rule
n o
h0|T Aa b
(x)A (y) |0i = Z0[J ]
iJ a(x) iJ b(y) J =0
k k
ab 1
iD (k) = iab g + (1 ) . (489)
k2 + i k2
Referring back to Eqs. (183) and (191), we had (for the case of
G = i/ m, but the result is clearly general)
Z i
1
Z
4
h
Z0 (, ) = [d ][d ] exp i d x (x)G (x) + (x) (x) + (x) (x)
N
Z
4 0 4 0 0 1 0 0 0
= exp i d x d y (x )G (x y ) (y ) . (491)
where
2 accb(x y) = ab 4(x y) , (493)
the solution to which is
d4k eik(xy)
Z
ab(x y) = ab . (494)
(2)4 k2 + i
where the explicit sign after the 2nd equality comes from anticommuting
c
the b(y) past the in taking the derivative of the exponential argument.
The above leads to the Feynman rule for the ghost field propagator in
momentum space:
ab i
i (k) = 2 ab . (496)
k + i
Note that the sign of this propagator is different in some treatments
but it is actually physically irrelevant so long as the relative sign of
the propagator and the ghost ghost gluon vertex remains unchanged.
This is because any Feynman graph computation will always contain one
propagator for every ghost-ghost-gluon vertex. The propagator sign and
the ghost-ghost-gluon vertex sign are simultaneously changed by changing
the (irrelevant) sign of the response function Mfab.
exp iSI , , a (497)
iJ a i a i
In SI we see:
1. a 3-gluon (here I use the QCD language) interaction;
2. a 4-gluon interaction;
3. a gluon-ghost-antighost interaction.
Since we set J = 0 at the end, this means that each of the J
4 0 0 0 t0 0
Z Z Z
4 def 4 x t
iSI Z0 [J ] d xgc i d y2 D dt (x y2 )J (y2 ) i d y2 D 0 (x y2 )J 0 (y2 )
et
" #
4 00 00 00 t00 00
Z
i d y2 D 00 (x y2 )J 00 (y2 ) exp . . . . (503)
ft
We now follow with the operation of Eq. (502). There are 3! = 6 ways
of performing the three final J s. We examine in detail one of the
terms and develop the others by symmetry. We focus on the term in
4 0 0 bt0
Z
0 4 0
d y2 D 0 (x y2 ) (y2 z ) g 0
et
4 00 00 ct00
Z
00 4 00
d y2 D 00 (x y2 ) (y2 w) g 00
ft
Z
def 4 x
= gc d x D da (x y ) D (x z )D (x w) .(505)
eb fc
a b c
h0|T {A (y )A (z )A (w)}|0i
4 4 def 3
h ih ih i
3 (2 ) (p + q + r )(gc )(i) (+ip ) iD da (p) iD (q ) iD (r )
eb fc
0 0 0 def
[iDad (p)][iD (q )][iD (r )] 0 0 0 (p, q, r ). (507)
be cf
from which we find the required form of for the particular one of
the 6 terms on which we are focusing.
def
0
0 0 (p, q, r) = gc def
(2) 4 4
(p + q + r)g 00 p0 . (508)
.
1
$ *
#%'"$&()
"!
0
!
. /
+ ( , -*
$ !
A
B
2 30 1 H0 1JI
4 . 45 4768:9 ;<9>=@? !DC EGF E 4 . 4 67=LK ;M? 4ON E A4P 4 68QK =@? . 4ON E . 4RA4 6;LK 8%? 4RS
1 .
R 0
abc
(p, q, r) = gc
abc
(2)4 4(p + q + r)gp . (512)
If external gluons are attached to this vertex, one would simply multiply
by the appropriate (p), . . ..
Next, let me return for a moment as to why I said that p was the
outgoing momentum associated with the a, field. This is because
of the way I defined Dda(x y) in terms of eip(xy). The e+ipx
goes with an a(p) in the standard 2nd-quantized field expansion,
corresponding to creation (i.e. which means outgoing) particle at the
interaction vertex location defined by x.
Now, we must generate the other 5 terms that I said we could get by
symmetry.
(a) First, we note that every term must be proportional to cabc (or some
permutation thereof that we can always write as cabc).
The group indices can only appear here. This cabc will be multiplied
by some expression involving the momenta and Lorentz indices of
the gauge particles.
(b) Next, since the gauge bosons are bosons, the vertex should be
symmetric under interchange of the gauge bosons (gluons in SU(3)
[g(p r) + g (r q) + g (q p) ] . (513)
One can check this structure for the required antisymmetries. For
example, under a, , p b, , q, the 1st term turns into the 2nd
term, and vice versa, while the 3rd term is antisymmetric on its own
when p q.
Thus, our final expression for the three-gluon (or more generally three-
gauge-boson vertex) is
Mabc
(p, q, r) = gc abc
[g(p r) + g (r q) + g (q p) ] ,
(515)
where the , , are to be contracted with the s (p), s 0 (q), s00 (r)
for the polarization vectors of the helicity states of the outgoing gluons
(indicated by the s, s0, s00 subscripts on the s), or used internally with
the , , indices connected to matching Lorentz indices of internal
propagators.
An aside on conventions and comparisons between these notes and
various books:
Some authors (like Ryder and Peskin) define a M0 that is related to
our M by iM0 = M. In Mandl-Shaw box normalization,
Y 1/2
1 X
h. . . |S| . . .i = 1 + 4 4
(2) ( pi)iM0 . (516)
2V E i
In this convention, our Feynman rules are for iM0. Ryder, Eq. (6.175),
and Peskin, Eq. (4.73), use continuum/covariant normalization, for
to
hp, a, s; q, b, s0; r, c, s00|0i , (519)
where I have given explicit helicity choices s, s0, s00 to the outgoing
gauge bosons.
For this, we need to recall some results from last quarter. We write,
with the appropriate trivial generalization to include color,
with
!1/2
a+
X 1
A = r (~
k)aa
r (~
k)e ikx
, (521)
2V ~k
r~
k
ab g
D (p) = ab . (524)
p2
d3 ~
x
Z
aa ~ r (~ ikx
ix0 Aa
r (k) = k) p e (x)
2V ~k
d3 ~
x
Z
~
aa
r (k) = r (~
k) p e ikx
ix0 Aa
(x) , (525)
2V ~k
where Ax0 B A(x0 B)(x0 A)B and the explicit sign is because
of the r (~
k) s(~
k) = rs for physical (transverse) polarizations.
Armed with the above identities, consider the vertex, including helicity
assignments for the outgoing gluons, and a notation out to remind us
that the annihilation operators appearing are those for outgoing gluons:
and focus on just one of the as for the moment, say acs00 (r). We write
(527)
where the is for w0 and where w0 +ei is the appropriate
w0
thing since we have stated all the gluons are in the outgoing state. The
imaginary component is the same one required to isolate the vacuum
at a very late time.
Related to the above time limits, you should be asking yourself why
d3 w
~ irw c
Z
0
= lim out hp, s, a; q, s , b|()s00 (~
r) p e iw0 A (w)|0iin
w0 e i 2V ~
r
d4 w irw c
d
Z
0
+ out hp, s, a; q, s , b|() 00 (~
r) p e iw0 A (w) |0iin
s 2V ~
r dw 0
0 c in
= out hp, s, a; q, s , b|[as00 (r )] |0iin
d4 w
Z
0 2 irw c irw 2 c
h i
+ out hp, s, a; q, s , b|() 00 (~
r) i (w e )A (w) + e (w A (w)) |0iin
0 0
p
s 2V ~
r
d4 w
Z
0 ~ 2 eirw )Ac (w) eirw ( 2 Ac (w)) |0i
h i
= 0 + out hp, s, a; q, s , b| 00 (~
r) p i ( w w0 in
s 2V ~
r
d4 w h
Z
0 irw 2 ~ 2 ]Ac (w) |0i
i
= + out hp, s, a; q, s , b| 00 (~
r) (i)e [ w w in
0
p
s 2V ~
r
d4 w
Z
irw 2 0 c
= i 00 (~
r) p e w out hp, s, a; q, s , b|A (w)|0iin , (531)
s 2V ~
r
0 c
out hp, s, a; q, s , b|A (w)|0iin
b out c
= out hp, s, a|[as0 (q )] A (w)|0iin
d3 z
~ iqz b
Z
c
= lim out hp, s, a|()s0 (q
~) q e iz0 A (z )A (w)|0iin
z0 +ei 2V q
~
d3 z
~ iqz
Z
b c
= lim out hp, s, a|()s0 (q
~) q e iz0 T {A (z )A (w)}|0iin
z0 +ei 2V q~
d3 z
~ iqz
Z
b c
= lim out hp, s, a|()s0 (q
~) q e iz0 T {A (z )A (w)}|0iin
z0 ei 2V q
~
d4 z iqz
d
Z
b c
+ out hp, s, a|() 0 (q
~) q e iz0 T {A (z )A (w)} |0iin
s 2V q dz0
~
d3 z
~ iqz c
Z
b
= lim out hp, s, a|()s0 (q
~) q e iz0 A (w)A (z )|0iin
z0 ei 2V q
~
d4 z
Z
2 iqz b c iqz 2 b c
h i
+ out hp, s, a|() 0 (q
~) i (z e )T {A (z )A (w)} + e z T {A (z )A (w)} |0iin
0 0
q
s 2V q
~
c b in
= out hp, s, a|A (w)[as0 (q )] |0iin
d4 z
Z
iqz 2 2 b c
h i
= 0 + out hp, s, a| 0 (q
~) i ()e ~
z z T {A (z )A (w)} |0iin
0
q
s 2V q
~
d4 z
Z
iqz 2 b c
= i 0 (q
~) q e z out hp, s, a|T {A (z )A (w)}|0iin . (532)
s 2V q
~
The most important new manipulation above was the insertion of the
time-ordered product. This was inserted so that when we converted
from z0 + to z0 the Ab (z) field moved all the way
over to being next to |0iin so that in the next step we had [abs0 (q)]in
sitting next to |0iin giving 0. Without the time ordering instruction
the [abs0 (q)]in operator would have been operating on Ac (w)|0iin, and
we could not say that it was annihilating the incoming vacuum.
At this point, we would continue in similar vein to reduce in the
[aas (p)]
out
operator. Following the same procedure, we would get the
analogous result involving the appropriate projection operator operating
on
a b c
out h0|T {A (y)A (z)A (w)}|0iin . (533)
Altogether, we get
0 00
out hp, a, s; q, b, s ; r, c, s |0iin
(534)
We now insert
a b c
h0|T {A (y )A (z )A (w)}|0i
d4 p0 ip0 y d4 q 0 iq 0 z d4 r 0 ir 0 w 0 0 0 0 0 0 def
Z Z Z
0 0 0
= e e e [iDad (p )][iD (q )][iD (r )] 0 0 0 (p , q , r ) ,
(2 )4 (2 )4 (2 ) 4 be cf
(535)
1 0
= q s ( p
~)g ad . (536)
2V p
~
2
Notice the cancellation of the propagator p0 from the result of the
0
y2 differentiation of eip y . Analogous results apply for the other two
projections, yielding 1
0 00 4 4
out hp, a, s; q, b, s ; r, c, s |0iin q q p (2 ) (p + q + r )M
2V p
~ 2V q
~ 2V ~
r
1 0 1 0
= s (p
~)g ad q 0 (q
~)g be
q
2V p 2V q s
~ ~
" #
1 0 4 4 def
p 00 (~
r )g cf (2 ) (p + q + r )gc [g 0 0 (p r ) 0 + . . .]
r s
2V ~
1 4 4 abc
= q q s (p
~) 0 (q
~) 00 (~
r )(2 ) (p + q + r )gc [g (p r ) + . . .] (537)
2V p 2V q
p
2V ~ s s
~ ~ r
Mabc
(p, q, r) = gc
abc
[g(p r) + g (r q) + g (q p) ] .
(538)
d3 ~
x ikx
Z
a out a out
[ar (~
k)] = r (~
k) q e ix0 [A (x)]
2V ~
k
d3 ~
x ikx 1
Z
a
= r (~
k) lim q e i x0 [ A (x )]( Z) (539)
x0 ei 2V ~
k
where Aa a
(x) is the fully interacting A field, whereas [A (x)]
out
is a free particle-like A field
with full one-particle normalization. The interacting A field in the very late time limit (or very early
time limit) does not have full one-particle normalization content and so we must
multiply by a factor
to boost up its normalization. This relative factor is by convention called Z with Z < 1 (after
renormalization or cut off). Another subtlety is that
out
lim A = ZA (540)
x0 ei
cannot be viewed as a statement about the operators, but only as a statement about the operators
when applied to out h0|.
In addition to the above ( Z )1 factor for each external particle, there is an additional Z factor for
each external particle, coming from the sum of so-called one-particle reducible diagrams inserted into
the external particle leg.
We have no time to get into all this. You will encounter these issues in a full discussion of the
reduction formalism.
0 0 0 0
4 0 4 0 a0 0 0 b0 0 0 c0 d0 0 0 d0 0
i
Z Z
4 0 4 0 c
exp d x d y J 0 (x )D 0 0 (x y )J 0 (y ) exp i d x d y (x ) (x y ) (y )
2 a b
!
ed00 d00
Z
Z
4 def 4
3 i d x gc d y1 (x y1 ) (y1 )
id (x) iJ f (x)
0 0 0 0
4 0 4 0 a0 0 0 b0 0 0 c0 d0 0 0 d0 0
i
Z Z
4 0 4 0 c
exp d x d y J 0 (x )D 0 0 (x y )J 0 (y ) exp i d x d y (x ) (x y ) (y )
2 a b
!
00 b00 ed00 d00
Z Z
Z
4 def 4 4
3 i d x gc d y2 D 00 (x y2 )J 00 (y2 ) d y1 (x y1 ) (y1 )
id (x) fb
0 0 0 0
4 0 4 0 a0 0 0 b0 0 0 c0 d0 0 0 d0 0
i
Z Z
4 0 4 0 c
exp d x d y J 0 (x )D 0 0 (x y )J 0 (y ) exp i d x d y (x ) (x y ) (y )
2 a b
ed00 d00
Z
4
d y1 (x y1 ) (y1 )
(543)
where there was a ()3 that arose when id(x) had to go past
two Grassmann objects to get to the object it acted on and then the
Grassmann quantity brought down from the exponential was moved
back to the left past a single Grassmann object.
So, now we want
" #
1 1 1
b a (z)
(544)
i (w) i J i c(y)
c00
(
c00 d 00 b00
Z Z Z
4 4 def 4
i d x x d x1 (x1 ) (x1 x) gc d y2 D 00 (x y2 )J 00 (y2 )
fb
)
ed00 d00
Z
4
d y1 (x y1 ) (y1 )
1
1
( Z
Z
c00 00 d
4 4 c
=
a (z ) i d x x d x1 (x1 ) (x1 x)
i b (w) i J
)
def
Z
4 00 b 00 h
ec
i
gc d y2 D 00 (x y2 )J 00 (y2 ) i (x y )
fb
(
c00
)
c00 d
Z
1
Z
4 4 def ec
n oh i
= i d x x d x1 (x1 ) (x1 x) gc iD (x z ) i (x y )
i b (w) fa
( Z )
4 bd def ec
h i n oh i
= i d x x i (w x) gc iD (x z ) i (x y )
fa
( Z Z Z
4 4 ip(xw) bd def 4 ik(xz )
= i d x x i d pe (p) gc i d ke D (k)
fa
Z )
4 iq(xy ) ec
i d qe (q )
(545)
Mabc
(k, p, q) = gc bca
p = gc abc
p , (547)
If you compare closely to Ryder, you will see that I think he has the
wrong sign for this vertex. In general, my results agree with those of
Bailin and Love (after accounting for the fact that they adopted the
opposite convention for the sign of g!).
d
X 1 h
d ~0 il0 x d ~0 il0 x
i
c (x) = p d (l )e + b (l )e
~
2V E~l0
l0
(548)
X 1 h
f
i
Af(x) = r )aft (~
t(~ r )eirx + t
(~ r )e+irx
r )at (~ .
t,~
r
2V ~r
(549)
To compute the Feynman vertex, we compute (s is the gluon helicity
1 ilx+il0 x+irx
Z
4 X X
= i d x q q p e
~ 0 t 2V E~ 2V E~ 0 2V ~
r
l,~
l ,~r l l
" #
a b
0 d 0 def t f e c
h0|as (~ ~) (il b (~
k )b ( p l )gc r )at (~
(~ r )b (~ ~) |0i
l) b ( q
1 ilx+il0 x+irx
Z
4 X X
= d x q q p e
2V E~ 2V E~ 0 2V ~
r
~
l,~ r t
l 0 ,~ l l
i
t 0 def
h
r )h0| st af ~
(~ bd ~ 0 l ce ~ gc |0i
~
rk p
~l l~
q
1 4 4 s bca
= q q q (2 ) (p + k q ) (~
k)gc p
2V Eq
~ 2V Ep
~ 2 V ~k
1 4 4 s
(2 ) (p + k q ) (~
k )M (550)
abc
q q q
2V Eq
~ 2V Ep
~ 2V ~
k
from which we obtain (where c belongs to incoming ghost of momentum
q, b to outgoing ghost of momentum p, and a to outgoing gluon of
momentum k) a result with opposite sign to Ryder (after careful index
mapping):
M abc = gc abc
p . (551)
"!
# $
$ !
# $
% !&
'
(554)
We first note that after using the ...s to perform the p0, q 0, r 0, s0
sums, we develop the integral
Z
d4xeix(p+q+r+s) = (2)4 4(p + q + r + s) , (559)
i 1 (~ 3 2 4
g 2ceabcecd
p) (~
r ) (~
q ) (~
s)
4
i 2 eab ecd 1 2 3 (~ 4
= g c c g g (~ p) q )
(~ r ) (~
s) (560)
4
As usual, in the 2nd line, I have restated the Lorentz dot products
of the s using the metric tensor. This is always a necessary step in
extracting the vertex function as defined in Eq. (552). Removing the
s, gives the first entry in our list of contraction contributions to the
vertex function:
0 0 i
a f, b g, c f , d g : g 2ceabcecdg g (561)
4
So, now where do the other 3 contributions that take exactly the same
0 0 i
a g, b f, c g , d f : g 2ceabcecdg g .(562)
4
Next, we can note that we get another two completely equivalent terms
by interchanging the roles of f, g f 0, g 0 note that this leaves the
AA contraction unaltered and the A A contraction unaltered in
each case:
0 0 i
a f , b g , c f, d g : g 2ceabcecdg g (563)
4
i 2 eab ecd
a g 0, b f 0, c g, d f : g c c g g .(564)
4
So, we have the required 4 identical terms to cancel the 1/4 factor. We
represent these 4 terms by the first sequence of contractions, writing
0 0 2 eab ecd
a f, b g, c f , d g + equiv. : ig c c g g . (565)
Hopefully, the game is clear now and we can write the complete
list without further ado. First, we give the forms obtained by the
permutations of the (, a), (, b), (, c), (, d) directly in the manner
that we have illustrated:
0 0 2 eab ecd
a f, b g, c f , d g + equiv. : ig c c g g (568)
0 0 2 eba ecd
(b f, a g, c f , d g + equiv.) : ig c c g g (569)
0 0 2 eac ebd
a f, c g, b f , d g + equiv. : ig c c g g (570)
0 0 2 eac edb
a f, c g, d f , b g + equiv. : ig c c g g (571)
0 0 2 ead ecb
a f, d g, c f , b g + equiv. : ig c c g g (572)
0 0 2 ead ebc
a f, d g, b f , c g + equiv. : ig c c g g . (573)
abcd 2 eab ecd
M (p, q, r, s) = ig c c g g g g
eac edb
+c c g g g g
ead ebc
+c c g g g g , (574)
ig 2ceabcecdg g (575)
For this, we need to introduce the fermion part of the action. Recall that
A
LF = [i (D)AB mAB ] B ,
with (D)AB = AB igLa a
AB .
A (578)
Going through the usual game of first getting the Z0 we would end up
with the obvious close analogue of the QED result:
Z
Z0(, ) = exp i d 4 x 0 d4 y 0 C
(x 0
)[S F (x 0
y 0
)] CD D
(y 0
) .
(579)
(582)
4 0 4 0 a0 0 0 0 0 0 b0 0
i
Z
exp d x d y J 0 (x )D 0 0 (x y )J 0 (y )
2 a b
Z
4 00 4 00 A 00 00 00 B 00
exp i d x d y (x )SF (x y )AB (y )
A0
Z Z
c 4 0 3 4 0
3 ig LCD d x () d x1 0 (x1 )[SF (x1 x )] 0 A0 C
00 b00 B0
Z Z
4 0 4 0
d y2 D (x y2 )J 00 (y2 ) d y1 [SF (x y1 )] 0 DB 0 0 (y1 )
cb00
exp{. . .} exp{. . .}
(583)
where, we will not be needing the exponentials anymore. So the next
stage is:
! ! !
1 1 1
i A (x) a
i J (y ) i B (z )
A0
Z Z
c 4 0 3 4 0
ig LCD d x ( ) d x1 0 (x1 )[SF (x1 x )] 0 A0 C
00 b00 B0
Z Z
4 0 4 0
d y2 D (x y2 )J 00 (y2 ) d y1 [SF (x y1 )] 0 DB 0 0 (y1 )
cb00
1
1 0 0
D ca (x y ) ( )()[SF (x z )] 0 DB 0 B 0 B 0
i i
Z
c 4 0 0
h i
= ig LCD d x iAC [SF (x x )]
0 0
n oh i
iD (x y )ca i[SF (x z )] DB
Z
4 0 0 c 0 0
h i h i
= ig d x [iSF (x x )] AC LCD [iSF (x z )] DB [iD ca (x y )]
4 0 ip0 (xx0 )
Z Z
4 0 0 c
h i
= ig d x d p e SF (p ) LCD
AC
4 ip(x0 z ) 4 ik(x0 y )
Z Z
d pe SF (p) DB d ke D ca (k)
c 4 4 0
igLCD (2 ) (p + k p) external propagators and trivial plane wave factors
a 4 4 0
igLAB (2 ) (p + k p) after standard relabeling . (584)
4 4 0
So, our Feynman rule for M (removing (2) (p +kp) and 1/ 2V E
factors that would appear in the reduction formalism approach) is:
[M
AB
a
] (p, k, p0
) = igLa
AB .
(585)
Figure 7: Propagator rules for the gluon, the ghost and a fundamental representation
fermion. a, b, c = 1, . . . , (N 2 1) are the gauge group indices (for SU (N )),
A, B = 1, . . . , N are fundamental fermion representation indices. For the gluon
propagator, we have hwritten the simple =i1 Feynman gauge form: more generally, the
k k
propagator contains g + k2 (1 ) . In the fermion propagator case, , are
Dirac indices.
-
.
/
0 #
23
.
/7 8 : 2
54 6 1 9 ;
1 8
-
=
6 ?>
@
@ < < A
0 =0
@
@ < ; <
0=
@
@ < ! < B,
< =
k1 + k 2
p1 + k 1 p1 + k 2
g k2 b q p2 B g k2 b q p2 B g k2 b q p2 B
t-channel s-channel u-channel
Figure 9: The t, s and u channel Feynman diagrams for gg qq.
For the two gluons, we take that with momentum k1 to have color a
and Lorentz index , and the 2nd to have color b and Lorentz index .
The outgoing quark has color label A and the outgoing antiquark has
color label B. The Feynman amplitudes for the two fermion exchange
diagrams are:
2 a iCD b
Mt+u (k1 ) (k2 ) = (ig ) uA (p1 ) LAC LDB
p/2 k/2 m
!
b i CD a
+ LAC LDB vB (p2 ) (k1 ) (k2 ) .
k/2 + p/1 m
(586)
2 a b
= (ig ) u(p1 ) i [L , L ] v (p2 )1
2 c abc
= g u(p1 ) L v (p2 )1 f (587)
where we went from the 2nd line to the 3rd line by using
p1 m) = 0 ,
u(p1)( / and (/
p2 + m)v(p2) = 0. (588)
The final result is non-zero because the color matrices do not commute,
unlike the abelian case where one has (effectively) [I, I] = 0. It is the
gluon exchange diagram that will cancel the above non-zero contribution.
It has exactly the right structure as it involves the color structure factor
f abc (labeling the exchange gluon with color index c) and also a gluon
quark antiquark vertex which will be assigned Lc.
Ms (k1 ) (k2 )
ig0
0 c
acb
= igu(p1 ) L v (p2 ) gf g ( k1 k2 ) + g ( k2 k3 )
2
k3
+g (k3 k1 ) (k1 ) (k2 )
2 k2 c i acb
= igu(p1 ) L v (p2 ) gf [k2 (k1 k2 ) + k2 (k2 k3 ) + g (k3 k1 ) k2 ] (k1 )
2
k3
k2 k3 k1 c i acb 2 2
= igu(p1 ) L v (p2 ) gf [g k3 + g k1 + k3 k3 + k1 k1 ] ( k1 )
2
k3
2 =0
k1 1 =0 , k1 c i acb 2
= igu(p1 ) L v (p2 ) gf [g k3 + k3 k3 ] (k1 )
2
k3
In proving the Ward identity above, we had to assume that the 2nd gauge
boson was transverse (k1 1 = 0).
Should we have expected that this would come out of the argument
rather than having to be an input? In QED, the Feynman diagrams
predict that the photons produced by e+e annihilation are transverse.
Amplitudes for producing the two other polarization states cancelled one
another. Recall:
If we go to a Lorentz frame where k = (k, 0, 0, k) has only a direction
3 vector component, then, in the Lorentz gauge we know that 1(~ k) =
(0, 1, 0, 0) and 2(~
k) = (0, 0, 1, 0), i.e. 11 = 1 and 22 = 1, all others
zero. So,
X = |M1|2 + |M2|2 , (590)
where the subscripts are the Lorentz indices. The Ward identity reduces
to
k(M0 + M3) = 0 , M0 = M3 . (591)
i(T T ) = T T . (594)
Take the matrix element between two-particle states, hp1p2| and |k1k2i,
and use the definition of M that has an extra i compared to what we
have been employing (which is more convenient for the moment):
1 1 1 1 4 4
hp1 p2 |iT |k1 k2 i = q q q q (2 ) (p1 + p2 k1 k2 )iM
f (k k p p ) .
1 2 1 2
2V Ep 2V Ep 2V E~ 2V E~
~1 ~2 k1 k2
(595)
The left hand side of Eq. (594) becomes
1 1 1 1 4 4
p p q q (2 ) (p1 + p2 k1 k2)
2V Ep~1 2V Ep~2 2V E~k 2V E~k
1 2
h i
iM(k1k2 p1p2) + iM (p1p2 k1k2) .
f f (596)
Let us now take the forward case of p1 = k1 and p2 = k2. In this case,
the above relation can be abbreviated as
XZ
2ImM
f (k 1 k2 k1 k 2 ) = f (k 1 k2 f )M
df M f (k1 k 2 f )
f
= 4Ecmpcmtot(k1k2 anything)
small masses
2stot(k1k2 anything) . (599)
f0 (k )+ (k ) c i acb 1
iM s 1 2 3 igv (p1 ) L u(p2 ) gf [k1 k1 ] (k1 )
2
k3 ~
2|k2 |
c i acb |~
k |
= igv (p1 ) L u(p2 ) gf k1 1 . (603)
2
k3 |~
k2 |
Of course, for tree-level processes we can just simply say that the
only external physical gluon states we allow are those with transverse
polarizations.
g k1 a c k1 a
q p1 A q p1 A q p1 A
Figure 10: The gluon-loop with two s-channel exchanges and the ghost
loop. In the gluon-loop case, where we show an s-channel exchange, one
should actually sum over the s, t and u-channel diagrams.
For the gluon loop, if we replace the gluon propagator g s by the sum
over four polarizations, Eq. (602), it appears that all the polarizations,
including the unphysical ones, should be included for the gauge bosons
in the sum over the qq gg amplitudes absolute squared. It is
the qq (ghost loop) qq diagram that must be added to the
one-loop qq (gluon loop) qq amplitude computation. The
ghost loop cancels off the loop contributions coming from the unphysical
polarizations hidden in the g s and only transverse polarizations for the
final state gs need be included in summing over the qq gg squared
amplitudes.
f (k1) (k2) ,
iM (605)
where, of course, the k1 and k2 are the same since we are examining the
situation where the initial state of the gg qq process is the same as
the final state of the qq gg process.
The Cutkosky rules state that4 DiscM f for the qq (gluon
f = 2iIm M
loop) qq diagram is obtained by replacing the cut gauge boson
propagators with momentum ki by
where the + means keep only the positive energy solution (i.e. (ki0))
4
Here, you should think of Mf as an analytic function with a branch cut across which there is a discontinuity. The
branch cut begins when the incoming cm energy is sufficient for the process to take place. A sketchy proof of the
Cutkosky rules will be given when renormalization is discussed.
d 4 k1 d 4 k2
Z
4 4 2 2 2
4 4
(2) (P k 1 k 2 )(2i) (k 1 )(k 2 )
(2) (2)
d3~
k1 d3~k2
Z
2 4 4
= (i) 3 3
(2) (P k1 k2)
(2) 2|k1| (2) 2|k2|
Z
= (i)2 d (607)
which is, aside from the (i)2, exactly the form of the final state phase
space for the gg final state of qq gg.
f T (k1)+ (k2) = 0
iM (609)
f0 5
since +
(k2 ) k2 . Precisely the same result applies also for iM .
5
This material is similar to that in Peskins Chapt. 16, but at this point I start disagreeing with him. In particular,
it is always +
(k) that is proportional to k , regardless of whether we are talking about k1 or k2 .
(k 1 ) +
(k 2 )
(k 1 ) +
(k 2 ) + +
(k 1 )
(k 2 ) +
(k 1 )
(k2 ) (611)
i
iMghost = igv(p2)L u(p1) 2 gf abck1
f 0 c
(614)
k3
c0 i ac0 b
0
iMghost = igu(p1)0 L v(p2) 2 gf
f k2 (615)
k3
0
for the gg qq direction. In the latter term, we replace k2
0 0
(p1 + p2 k1) , use the fact that the (p1 + p2) portion gives
zero by virtue of the Dirac equations for v(p2) and u(p1) and use
ac0 b abc0 f0
f = f . The product (iM ghost )(iMghost ) is then of exactly
f
the same form as the first of the two gluon-loop terms in Eq. (612);
the factor of 21 in the latter is compensated by the fact that the 2nd
product term simply doubles the first product term. It would seem
that the ghost loop simply doubles the net result from the gluon loop?
Is there some reason for which they cancel? The answer is YES. They
cancel because we still must include the () sign for the ghost loop.
So, we have verified that the Faddeev-Popov ghosts serve to cancel
the timelike and longitudinal polarization states of the gauge bosons in
the context of situations (such as unitarity) where the gauge bosons of
internal loops are placed on shell.
Our Feynman rules for M will then take the factorized form:
1. () sign for closed fermion or ghost loops;
R d 4 ki
2. (2)4 for each independent loop i;
3. Other internal loop momenta are determined by momentum conservation
at each vertex;
4. Symmetry factors, such as a factor of 1/2 for gluon loop insertion into
gluon propagator;
5. Times, the following graphical rules.
% %
& '
"!$#
(
% %
&
) *+
-,/. & 20 131
46587:9; ) * =<?>
Figure 11: Propagator rules for the gluon, the ghost and a fundamental
representation fermion in the Feynman/color factorized form.
! ,+-
" # $ &%(' *)
.+
/
01 2+-
' 2)
+
45
/
198 :4 '
<; = 3
76
3 =
.+
A@: 2+-C *) C<D?
> > B ' 2 )
2
C *) C<+?
E : > > ' 2 )
E > > ' 2) 2 ?2C *) C<D+ %
> ? :
Figure 12: Factorized forms for Feynman vertex rules: 3-gluon, ghostghost+gluon,
fermionfermion+gluon and 4-gluon. We will draw a separate Feynman/momentum diagram times a
color diagram for each of the expressions given; in particular, the 4-gluon vertex will be reduced to the
sum of three pieces, each represented in this factorized form. Note that I have set up a certain clock-wise
convention for the icabc factors; in order to do this for the three 4-gluon terms, I changed the 2nd and 3rd
metric tensor combination signs.
Figure 14: The color diagrams for the three subdiagrams coming from the
4-gluon interaction. They have an s-, u- and t-channel type topology,
respectively, from the perspective of the original diagram.
We now need to learn a bit better how to deal with color algebra. There
are many different techniques. Here, I will discuss a technique developed
by Cvitanovic in PRD14, p1536.
It is a diagrammatically based technique, that once mastered allows
extremely rapid computation of color factors. We will focus on the
application to SU (N ) groups, and in particular our notation will be
adapted for QCD SU (3).
Cvitanovic uses slightly different conventions regarding index names than
i i
those I have employed above. In his notation a i for the L = 2 , and
A a for the fundamental fermion representation color index. I will try
to stick to our earlier notation.
So, we have
X 0 0
a b
[L , L ] = i cabc Lc (616)
c0
In words, this last result says (in QCD) that a gluon is almost a
quark-antiquark combination except that the singlet quark-antiquark
combination is subtracted. Finally,
X
LdBC LdCA = CF BA (623)
d
Graphically, this is
represented by:
N2 1 4
CF = = for N = 3 (QCD). (624)
2N 3
Another little example that illustrates how we will use these color
computations when it comes to analyzing the effects of loops is the
following result.
represented by:
The computation is an application of Eq. (619). Graphically we have:
where the (2)2 comes from using Eq. (619) twice, and the single (2)
comes from the fact that doing so actually yields 4 terms containing two
pairs of identical terms. Note: the arrows on the loops are in the same
direction for the 1st term and are in opposite directions for the 2nd term.
The first term inside the brackets is, using Eq. (620) (in graphical form),
the definition of CF and then Eq. (618) (in graphical form) for the
successive equalities:
J. Gunion 230B, 2nd Quarter of Field Theory 295
The second term inside the brackets is reduced using the graphical forms
of Eqs. (620) and (618) as follows:
= 12 [ N1 ]
= 12 [ N1 ]
= 21 [ 12 { N1 } N1 ( 21 )2 ]
= 12 [ 12 {(0)2 N1 ( 12 ) } N1 ( 21 )2 ]
1
= 4N
!
"
implying that CA = N .
CA
= 2.23 0.01(statistical) 0.14(systematic) . (626)
CF
PRuR 0 = PR /
k1uL 0 = /
k1PLuL 0 = /
k1uL 0 = uR 0 . (632)
1 2
0 p + ip
/
puR 0 1 0 1 +
p
uL(p) = =p / p
=p (634)
2p k0 p+ 0 p+ 0
1 0
That /
puL(p) = 0 is evident from /
p/p = p2 = 0 (for a massless particle).
That PLuL(p) = uL(p) is evident from PL /puR 0 = / pPRuR 0 = / puR 0.
However one arrives at these spinors (one could have just constructed
uR(p) and uL(p) explicitly by rotating uR 0 and uL 0 from the negative z
momentum direction to the general p direction), they are the right- and
left-handed spinors for a massless particle with momentum p as written
in the Weyl representation. The only difference at this point with GK is
that GK use the Dirac representation for these same spinors.
0 1 1 2 +
uL(p) = uL(p) = p 0, 0, (p + ip ), p (638)
p +
where we have given the Weyl form of 0. The spinor inner products
are defined by
s s
p+ 1 k+ 1
2 2
uR(k)uL(p) hk + |pi = +
( k ik ) +
( p ip ) (640)
k p
s s
p + k+ 1
1 2 2
uL ( k ) uR ( p ) hk |p+i = ( k + ik ) + ( p + ip ) , (641)
k+ p+
with the same result for |hk |p+i|2, implying that the inner products
can be thought of as the square roots of twice the dot product. Another
way of arriving at this same result illustrates some further manipulation
techniques:
2
|hk + |pi|
= hk + |pihk + |pi = hk + |pihp |k+i = hk + |(|pihp | + |p+ihp + |)|k+i
where for the 3rd equality we used Eq. (642) and for the 6th equality
we used the fact that PR|ki = 0, as is obvious from the definition of
negative helicity and also the explicit expression for |ki uL(k).
where
0 1 0 0
2 0
1 0 0 0
C = i =
0 0
(658)
0 1
0 0 1 0
hk | |pi
[ (k, p)] = . (660)
2hp |ki
+ 1 p+
[ (k, p)] = (0, 1, i, 0) 2 + 1 2)
k = standard+k ,
2 k (p + ip
(661)
and
1 p +
[ (k, p)] = (0, 1, i, 0) 2 + 1 2
k
= standard + 0
k ,
2 k (p ip )
(662)
where by standard we mean we are getting the results equivalent to
1
() = (0, 1, i, 0) (663)
2
hk + | |p+i hk + | |l+i
" #
+ + 1
[ (k, p)] [ (k, l)] =
2 hp |k+i hl |k+i
/ |p+i hp |k
1 hl |k / |l+i
=
2 hp |k+ihl |k+i
/ |p+i + hl | k|p
1 hl |k / +i
= using the C game
2 hp |k+ihl |k+i
hl |p+i
= 2k (664)
hp |k+ihl |k+i
which is precisely the expression for the gluon propagator in the so-called
light-like axial gauge with the light-like axial vector n chosen to be p.
where the Greek indices are Dirac indices. A proof following the Peskin
5.3 route appears in a short section following this section. If we now
sandwich the between the two + spinors u+(a) . . . u+(b) , we
obtain
u+ (a) ( PR ) u+ (b) ( PL ) = 2(PR ) u+ (b) u+ (a) (PL ) = 2u+ (b) u+ (a) .
(667)
In our bra-ket notation, this is written as
Note that this result also implies, using Eq. (656), that
3. Finally, from the Fierz identities Eqs. (670)-(673), we find the important
identity needed when an external polarization for a gluon is Lorentz
hk + | |p+iPL = 2|p+ihk + |
hk + | |p+iPR = 2|kihp |
hk | |piPL = 2|k+ihp + |
hk | |piPR = 2|pihk | . (677)
So, now just a few words about using this formalism in particular for
the gluons.
First, consider external gluons. The idea is to choose reference
momentum p in (k, p) that will simplify your calculation. Typically,
the simplification will occur if p is chosen equal to one of the external
fermion momenta. In this way, one or more diagrams can often be
zeroed. Of course, once you make a choice of p for computing one
1
uL(p1) uL(p2)[ (1 5)] = 2 [uR(p1)uR(p2)] . (681)
2
Clearly, the right hand side of the above equation must just be some
4 4 matrix M in Dirac space and so it can be written as a linear
combination of the standard 16 matrices (1, 5, , 5, 21 [, ]).
The only question is which of these actually appear on the left hand side.
1 1
5
M = (1 )V + (1 + 5)W , (682)
2 2
where in the last step we used Eq. (656). Meanwhile, we would compute
W as
1 1
W = Tr (1 + 5) (2uR(p1)uR(p2))
2 2
1
uL(p1) uL(p2)[ (1 + 5)] = 2 [uL(p2)uL(p1)] . (686)
2
We would prove this in just the same manner as above except that it
would be W that would survive and we would compute
1 1
W = Tr (1 + 5) (2uL(p2)uL(p1))
2 2
1 1
= Tr (1 5)(2uL(p2)uL(p1))
2 2
1
= Tr [ 2uL(p2)uL(p1))]
2
= uL(p1) uL(p2) , (687)
Applications
d 2
(e
R (k)e +
L (p)
R (k 0
)+
L (p0
)) = 2
(1 + cos )2 . (688)
d 4Ecm
g
M = u+(k0)iev(p0)i 2
v (p)ie u+(k)
(p + k)
e2
= i hk0 + ||p0+ihp + | |k+i
s
e2
= i 2hp + |k0ihp0 |k+i , (689)
s
where we used one of the Fierz identities, Eq. (675), for the last equality
above. From this, we obtain (using |ha + |bi|2 = 2a b)
2 e4 0 0
|M| = 4 2
(2p k )(2p k) . (690)
s
Now, we need to recall our kinematics
k = (E, 0, 0, E) , p = (E, 0, 0, E)
k0 = (E, E sin , 0, E cos )
p0 = (E, E sin , 0, E cos ) , (691)
1
2p k0 = 2p0 k = 2E 2(1 + cos ) = 2
Ecm (1 + cos ) (692)
2
so that
2 e4 1 4
|M| = 4 4
E (1
4 cm
+ cos )2 = e4(1 + cos )2 . (693)
Ecm
d 1 |~
k| 2
= 2 16 2 E
|M|
d 2Ecm cm
1 2
= |M| for E m, |~
k| = Ecm/2
64 2Ecm
2
e4
= (1 + cos )2
64 2Ecm
2
d
(e
R (k)e +
R (p)
R (k 0
) +
L (p 0
)) (695)
d
requires the amplitude
0 0 g
M = u+(k )iev(p )i v +(p)ie u+(k)
(p + k)2
e2
= i hk0 + ||p0+ihp | |k+i
s
= 0 (696)
So, now let us attempt a real QCD calculation. The process I will discuss
is gg qq. We wish to obtain the form of the cross section
d 1 1 X 2
= |M| (697)
dt 16s2 nspinncolor colors,spins
where, as indicated, we must average over initial spins and over initial
colors.
The result we will obtain when the quark masses can be neglected is
2s 1 u2 + t2 3 u2 + t2
d
= . (698)
dt s2 6 ut 8 s2
As you will see, even using the tricks of the massless limit, the calculation
is not easy.
M = A t C t + A s C s + Au C u (699)
First, let us take care of the color algebra. We have two initial gluons.
Each of these initial gluons can have N 2 1 = 8 different colors. Thus,
ncolor = (N 2 1)2 = 64.
1. The square of the t channel color factor produces, after summing over
the colors of the initial gluons and of the final quarks, a closed quark
loop with two gluon CF type attachments leading to |Ct|2 = CF2 N =
2 2
N 1 (N 2 1)2
N 2N = 4N = 16 3
.
Figure 19: The color factor computation for CtCu.
5. Finally, we have the color calculation for CtCs = CuCs, where the
latter sign comes from the fact that the gluons are crossed (in their
attachment to a closed quark loop) in the CuCs calculation compared
to the CtCs calculation.
The CtCs diagram has a closed quark loop to which 3 gluons attach.
Two of the gluons circle from left to right to attach to a 3-gluon vertex
So, now we turn to the Feynman helicity amplitudes, As,t,u(1, 2, s1, s2).
1(1)( /
At(1, 2, +, +) = hp1 + | / p1 + / 2(2)|p2i = 0 (701)
k1 ) /
where the final |p2i comes from remembering that v+(p2) = u(p2),
and the final result of 0 arises because we need to have an even number
of matrices between a bra and ket of opposite helicities.
A similar argument applies to the s and u channel diagrams for the
s1 = s2 = + helicity configurations, as well as to the s1 = s2 =
helicity configurations.
2. Next consider s1 = +, s2 = . Here we must work case by case in
1, 2.
Consider first 1 = 2 = +.
The t-channel diagram gives a numerator algebra:
1(1)( /
At(1, 2, +, ) = hp1 + | / p1 + / 2(2)|p2+i .
k1 ) / (702)
2(2)( /
Au(1, 2, +, ) = hp1 + | / p1 + / 1(1)|p2+i .
k2 ) / (704)
where
V 1 (1)
2 (2 ) [(k1 k2 ) g + (2k2 + k1 ) g (2k1 + k2 ) g ] .
(706)
1(1)( /
At(1, 2, +, ) = hp1 + | / p1 + / 2(2)|p2+i ,
k1 ) / (708)
|k1+ihp1 + | + |p1ihk1 |
hp1 + | /1 (k 1 , p1 )
= hp1 + | 2
...
(709)
which gives 0.
Similarly, for Au we will get zero if we choose p1 as the reference
momentum for 2().
As in the previous 1 = 2 = + case, we will also find that the
s-channel diagram is zero in the present 1 = 2 = case when both
1() and 2() are referenced to p1.
Note: In case you have not noticed, we have been very careful to
choose the same reference momentum (same gauge) for all diagrams
contributing to a given helicity configuration. We are free to change
the gauge when computing the diagrams for a different helicity
configuration. This is allowed, since each helicity configuration is an
independent physical observable.
Thus, by a somewhat clever choice of gauge, we have been able
to show that all helicity amplitudes other than A(+, , +, ) and
|k2+ihp1 + | + |p1ihk2 |
/2 (k 2 , p1 )
= 2
hp1 + |k2i
|k1ihp2 | + |p2+ihk1 + |
+
/1 (k 1 , p2 ) = + 2
hp2 |k1+i
hk2 | |p1i
2 (k2 , p1 ) =
2hp1 + |k2i
hk1 + | |p2+i
+
1 (k1 , p2 ) = + (710)
2hp2 |k1+i
The s-channel amplitude takes the form, now including all factors
carefully:
ig0
+
As = igs hp1 + | |p2 +i2 (k2 , p1 ) 1 (k1 , p2 )
( k1 + k2 ) 2
h i
(igs ) g0 (2k1 k2 ) + g (k1 k2 )0 + (2k1 + k2 ) g0
where, in the above, we used the explicit forms given for this case
earlier and noted that
hp1 + | /2
(k 2 , p1 )|p2 +i = hp1 + | /
1 (k1 , p2 )|p2 +i = 0 (714)
for the reference momentum choices made, implying that we only had
to keep the g term in the vertex component in the above reduction.
Now, we use k2 = (p1 + p2 + k1) and hp1 + | / p1 = /p2|p2+i = 0 to
rewrite
hp1 + |( /
k1 /
k2)|p2+i = hp1 + |(2 /
k1)|p2+i
= 2hp1 + |k1ihk1 |p2+i
= 2hp1 + |k1ihp2 |k1+i . (715)
Inserting this into Eq. (713), and canceling a common hp2 |k1+i in
and
Our two helicity amplitudes are then, bringing back in the color factors,
We now recall our color factors. Including the 1/64 for initial gluon color
1 3 3
Ct2 = Cu2 = , Cs2 = , CtCs = CuCs = . (723)
12 16 32
2 2 1 2 2
|M(+ +)| + |M( + +)| = |At (1, 2)| + |At (2, 1)|
12
3 2 2
+ |As (1, 2)| + |As (2, 1)| ReAt (1, 2)As (1, 2) ReAt (2, 1)As (2, 1) (724)
16
which exhibits the expected final symmetry under interchange of the two
initial gluons. (Note how the sign of CuCs made up for the extra
sign in the momentum-space calculation of As( + +).)
The final task is then to compute explicitly the various A products. This
2
is easy using |ha + |bi| = 2a b. We have
2k1 p12k2 p2
2 t
|At(1, 2)| = 4gs4 = 4gs 4
, (725)
2k1 p12p1 k2 u
Further,
At(1, 2)As (1, 2) = At (1, 2)As(1, 2) . (731)
u3 t3 t2 u2
" !#
2 2 4 1 t u 3
|M(+ +)| + |M( + +)| = 4gs + + + + + . (732)
12 u t 16 s2 t s2 u su st
t3 t2 1 1 t2
+ = t2(t + s) = t2(u) = , (733)
s2u su s2u s2u s2
and, similarly,
u3 u2 u2
+ = , (734)
s2t st s2
resulting in
1 X 2 1 2 2
|M(1 , 2 , s2 , s2 )| = 2 |M(+, , +, )| + |M(, +, +, )|
4 4
1 ,2 ,s1 ,s2
" #
2 2
4 1 t u 3 u + t
= gs + , (735)
6 u t 8 s2
d 1 1 X
M(1, 2, s2, s2)2
=
dt 16s2 4 , ,s ,s
1 2 1 2
(4s)2 1 t 3 u2 + t2
u
= 2
+
16s 6 u t 8 s2
2s 1 t 3 u2 + t2
u
= 2
+
s 6 u t 8 s2
2 2 2 2 2
s 1 t + u 3u + t
= , (736)
s2 6 tu 8 s2
s0 = t , u0 = u , t0 = s . (737)
We must also account for the change in color averaging. The color
averaging factor of 1/(N 2 1)2 = 1/64 for gg qq must be replaced
by 1/[N (N 2 1)] = 1/24 for gq gq. The ratio gives a factor of
8/3. Thus, up to a sign that I will discuss shortly, we have (dropping the
prime notation)
d gqgq 8 d ggqq
(s, t, u) = (t, s, u)
dt 3 dt
8 s 1 u2 + s2
2
3u + s 2 2
=
3 s2 6 su 8 t2
s u + s2
2 2
4u + s2 2
= . (738)
s2 t2 9 us
This expression is obviously < 0, and so there was some subtlety that
the above argument misses. To get the sign straight, we must return to
the underlying procedures and recognize an important subtlety.
In terms of the notation we have adopted here, this means that the
final outgoing antiquark momentum p2 is crossed to an incoming quark
with outgoing momentum p2, whereas (in our all outgoing notation)
we want this momentum to be +p2.
2
P
This means that we must also change the sign of colors,spins |M| at
the same time that we perform the s t interchange, yielding
2s 2 2 2 2
d u +s 4u + s
= + 2 . (741)
dt s t2 9 us