3

When we "expand" our classical fields, for example the Dirac field, in the standard way which we later go on to "quantise":

$$\psi(x,t)=\int d^3\tilde k \sum_{a=1,2}\left(b_a(k)u^a(k)e^{-ik\cdot x}+d_a^\dagger(k)v^a(k)e^{ik\cdot x}\right), \tag{1}$$

for what reason do we choose $e^{ip\cdot x}$ and $e^{-ip\cdot x}$ as our "basis" of functions. Is it possible to do this procedure with different functions, and if so, for what reason do we not (at least not in any standard introductory texts I have ever read) do so?

Charlie
  • 6,955
  • 1
    I think I remember in Quantum Mechanics by Messiah, he expands the EM field in spherical harmonics for calculating atomic transitions. This would be in volume 2, although they're usually published together. This is a standard graduate text, although it is pretty old now. – TEH Jan 01 '21 at 00:33

2 Answers2

3

This is done so that the $S$-matrix for the fields is manifestly invariant under translations. I will illustrate the case explicitly for a real scalar field $\phi$, with creation and annihilation operators $a^\dagger (\vec{p})$ and $a(\vec{p})$, respectively. Under translations $x \to x + b$, these transform as $$ a'(\vec{p}) = U(b) a(\vec{p}) U^{-1} (b) = e^{ip \cdot b} a(\vec{p}) $$ and $$ a'^{\dagger}(\vec{p}) = U(b) a^\dagger(\vec{p}) U^{-1} (b) = e^{-ip \cdot b} a^\dagger(\vec{p}) $$ where $p^0 = \sqrt{\vec{p}^2 + m^2}$, and $U$ is the unitary operator that implements the translation. The field is built out of creation and annihilation operators: $\phi(x) = \int d^3 \vec{p} \left( u(x,\vec{p}) a(\vec{p}) + v(x,\vec{p}) a^\dagger (\vec{p}) \right) $, where $u$ and $v$ are some functions which have yet to be determined.

The key point is that we require $\phi$ to be invariant under translations; that is, $$ \phi'(x) = U(b) \phi(x)U^{-1}(b) = \phi(x + b). $$ We impose this requirement so that a Hamiltonian constructed from $\phi$, and hence the corresponding $S$-matrix, will also be translationally invariant. From this $u$ and $v$ are determined, up to a function of $\vec{p}$: $$ u(x,\vec{p}) = e^{ip\cdot x} u(\vec{p}); \qquad v(x,\vec{p}) = e^{-ip\cdot x} v(\vec{p}). $$ The $x$-dependence is fixed by translational invariance.

This argument works for any fields, not just scalars. In the general case $u$ and $v$ will carry some indices corresponding to the different field components.

Reference: Weinberg, The quantum theory of fields, volume I, $\S$5.1,5.2. For the transformation of creation and annihilation operators under translations, see $\S$4.2.

  • I haven't gotten to reading Weinberg's book yet, it seems to be considered quite the holy grail for QFT. Just a quick follow up, it seems that we choose the $u(\vec p)$ and $v(\vec p)$ you gave above to be $=1$ in the real scalar field, is there a reason for this? – Charlie Jan 01 '21 at 11:41
  • Also on a slightly different note, when we say translationally "invariant" here do we mean covariant or genuinely invariant? It seems odd to me that our fields would be invariant under translation (i.e. look identical when translated). – Charlie Jan 01 '21 at 11:46
2

Is it possible to do this procedure with different functions...?

Yes.

...for what reason do we not (at least not in any standard introductory texts I have ever read) do so?

No fundamental reason. It's just convenient in some contexts, that's all. In flat spacetime, using solutions with translation symmetry is often convenient, but it's not necessary. The theory as a whole still has translation symmetry even if use mode functions that don't individually have translation symmetry.

I'll explain how to formulate things using generic mode functions in flat spacetime. This approach is also important for quantum field theory in curved spacetime, because the curved-spacetime version of the Dirac equation typically doesn't have plane-wave solutions.

Quantization with general mode functions

We want the field operators to satisfy these equal-time anticommutation relations: $$ \newcommand{\bfx}{\mathbf{x}} \newcommand{\bfy}{\mathbf{y}} \newcommand{\bfp}{\mathbf{p}} \newcommand{\dpsi}{\psi^\dagger} \big\{\psi_a(\bfx,t),\,\dpsi_b(\bfy,t)\big\} \propto\delta_{a,b}\delta^3(\bfx-\bfy). \tag{1} $$ We also want them to satisfy the Dirac equation $$ (i\gamma\partial+m)\psi(\bfx,t)=0. \tag{2} $$ The Dirac equation implicitly tells us how to express the field operators at times $t\neq 0$ in terms of the field operators at time $t=0$, so we need to check that the left-hand side of equation (1) really is independent of time. This consistency-check is a by-product of the following analysis.

Let $f_a(\bfx,t)$ and $g_a(\bfx,t)$ be ordinary complex-valued functions, where $a$ is a spinor index, and define $$ (f,g) \equiv \sum_a\int d^3x\ f_a^*(\bfx,t)g_a(\bfx,t). \tag{3} $$ If $f$ and $g$ are solutions of the Dirac equation, then $(f,g)$ is independent of time. To prove this, use the fact that equation (2) can be written as $i\dot f=hf$, where the differential operator $$ h=-\gamma^0\left(i\sum_{k>0}\gamma^k\partial_k+m\right) \tag{4} $$ satisfies $(g,hf)=(hg,f)$. This shows that $(f,g)$ is independent of time. Now let $M$ be a set of mode functions, solutions to the Dirac equation that are linearly independent and that satisfy the completeness relation $$ \sum_{f\in M} f_a(\bfx,t) (f,g) = g_a(\bfx,t) \tag{5} $$ and the non-overcompleteness relation $$ (f,g) = 0 \hspace{2cm} \text{if }f,g\in M\text{ and }f\neq g. \tag{6} $$ The set $M$ of mode functions is not necessarily countable, so the "sum" in (5) should be understood in a suitably general sense. We can use these mode functions to solve equations (1)-(2). Let $\psi(f)$ be a time-independent operator "indexed" by the mode functions $f\in M$, and impose the anticommutation relations $$ \big\{\psi(f),\,\dpsi(g)\big\} \propto (f,g). \tag{7} $$ The mode operators $\psi(f)$ are time-independent by definition. This does not contradict the fact that $f$ is time-dependent, because $f$ is being used here only as an "index" to distinguish one operator from another. Now equations (1)-(2) are automatically satisfied by $$ \psi_a(\bfx,t)=\sum_{f\in M} f_a(\bfx,t)\psi(f). \tag{8} $$ This generalizes the equation shown in the question. In hindsight, we can use equations (6) and (8) to derive the satisfying relationship $$ \psi(f)=\sum_a\int d^3x\ f_a^*(\bfx,t)\psi(\bfx,t). \tag{9} $$ As a consistency check, notice that the right-hand side is actually independent of time for the same reason that $(f,g)$ is independent of time.

Constructing a vacuum representation

To construct a Hilbert-space representation with a lowest-energy (vacuum) state, we can impose one more condition on the mode functions: each one of them should involve either only positive frequencies or only negative frequencies. This is desirable because the positive- and negative-frequency parts of a field operator (or any other Heisenberg-picture operator) act as energy-lowering and -raising operators, respectively, as explained in my answer to another question:

Write $M=M_+\cup M_-$, where $M_+$ and $M_-$ are the sets of positive- and negative-frequency mode functions, respectively. Define a state $|0\rangle$ by the conditions \begin{gather} \psi(f)|0\rangle = 0 \hspace{1cm} \text{for }f\in M_+ \\ \dpsi(f)|0\rangle = 0 \hspace{1cm} \text{for }f\in M_-. \tag{10} \end{gather} All other states obtained by acting on $|0\rangle$ with the other mode operators have higher energy than $|0\rangle$, so $|0\rangle$ is the vacuum state. We say that the other mode operators $\dpsi(f)$ with $f\in M_+$ and $\psi(f)$ with $f\in M_-$ create particle and antiparticles.

Recovering the version shown in the equation

To recover the version shown in the question, take the set of mode functions to be a complete set of single-frequency plane-wave solutions of the Dirac equation. To conform to tradition, write $b$ instead $\psi(f)$ when $f$ has positive frequency, and write $d^\dagger$ instead of $\psi(f)$ when $f$ has negative frequency. Equations (10) say that the operators $b$ and $d$ annihilate the vacuum state. Their adjoints $b^\dagger$ and $d^\dagger$ create particles and antiparticles.

Spontaneous particle creation in curved spacetime

The same idea of using generic mode functions also works in curved spacetime, where plane-wave solutions might not be available. In a time-dependent curved spacetime (or in the presence of other time-dependent background fields), the equation of motion may not preserve separate positive- and negative-frequency parts. (These can still be defined in asymptotic regions where spacetime is essentially flat.) In other words, a solution that has positive frequency initially may end up having a mix of positive and negative frequencies, and this can have interesting effects. We can choose the state to be annihilated by the mode-operators that have positive frequency initially, but then it won't be annihilated by the mode-operators that have positive frequency later. The result is that the initial vacuum state evolves into a non-vacuum state. This is how Hawking radiation is derived, using the equation of motion for the field in the time-dependent spacetime of a collapsing star.

  • 1
    I really appreciate your detailed answer, it's a bit advanced for me at the moment but hopefully it helps others who have a similar question. – Charlie Jan 01 '21 at 11:45