Post

Stochastic Quantization

Stochastic Quantization

Notes on Stochastic Quantization (Draft)

Abstract

These notes collect several essential ideas behind stochastic quantization approached from a numerical perspective. The intention is to serve as a guide to the literature for students and practitioners. Emphasis is given to connecting the stochastic quantization formalism to the disciplines of numerical relativity and quantum field theory on curved spacetimes.


Stochastic quantization, broadly viewed, includes several active research programs: Lattice field theory / Monte Carlo simulation, the sign problem in finite-density QCD, semiclassical and stochastic gravity. The bibliography here reflects that breadth. Spanning from the Martin–Siggia–Rose formalism of the 1970s through contemporary work on complex Langevin regularization, the notes serve as both an orientation to the conceptual landscape and a framework for computational projects that accompany the notes.

In the following, natural units $\hbar = c = 1$ are employed unless otherwise indicated. The metric signature is $(-,+,+,\ldots)$ for Lorentzian spacetimes. The Euclidean signature is $(+,+,+,\ldots)$ where explicitly stated. The stochastic or “fictitious” time is denoted $s$, to distinguish it from physical time $t$ or Euclidean time $\tau$.


1. Introduction

The standard routes to quantum field theory—canonical quantization and Feynman path integrals—are by now thoroughly developed. Why introduce another? The answer is partly practical and partly conceptual, and the balance between these motivations has shifted over the decades since Parisi and Wu’s foundational paper.

The practical motivation is gauge theory. In the path integral formulation, quantizing a gauge theory requires fixing the gauge redundancy to avoid overcounting equivalent configurations. The Faddeev–Popov procedure accomplishes this but introduces ghost fields, gauge-fixing terms, and the attendant complications of BRST symmetry. For non-Abelian theories, the procedure is further complicated by the Gribov ambiguity: the gauge-fixing surface may intersect gauge orbits more than once, so that the Faddeev–Popov determinant does not fully resolve the redundancy. Parisi and Wu observed that stochastic quantization avoids explicit gauge fixing entirely. The gauge degrees of freedom undergo a random walk along the gauge orbit, contributing nothing to gauge-invariant observables in the equilibrium limit.

The conceptual motivation is the link between quantum mechanics and statistical mechanics. Stochastic quantization makes this link dynamical: one literally constructs a stochastic process whose equilibrium distribution is the Euclidean path integral measure $e^{-S[\Phi]}$. This is not merely a formal trick. It provides an alternative derivation of the Feynman–Kac formula, offers a natural framework for non-equilibrium phenomena (via finite stochastic time), and—as we shall see—suggests connections to semiclassical gravity that are not obvious from the canonical or path integral perspectives alone.

The computational motivation is perhaps the most enduring. The Langevin equation that defines stochastic quantization is, at bottom, a stochastic partial differential equation (SPDE). Solving SPDEs numerically is a well-developed art, and stochastic quantization provides an alternative to the Metropolis–Hastings algorithm for sampling field configurations. In certain contexts—particularly when the action is complex-valued, as in finite-density QCD—the Langevin approach (suitably complexified) offers a route where importance sampling fails entirely.

The notes that follow develop these themes in roughly the order that a practitioner would encounter them. The basic formalism is first considered in flat Euclidean space prior to proof of equivalence via Fokker–Planck theory. From there consideration is expanded to gauge theories, then curved backgrounds, and finally the extensions to complex actions and gravity. The bibliography is organized to support this progression.

Historical context

The roots of stochastic quantization extend to before 1981 and Parisi and Wu. Quantum mechanics and stochastic processes were explored by Nelson (1966) in his “stochastic mechanics,” which attempted to derive the Schrödinger equation from a diffusion process in physical time. The Martin–Siggia–Rose (MSR) formalism (1973) [S1] provided a path integral representation of classical stochastic dynamics, and the theory of dynamic critical phenomena (Hohenberg and Halperin, 1977) [S2] established the technology of Langevin equations coupled to order parameter fields. Fox (1978) [S3] systematized the treatment of Gaussian stochastic processes in physics. The closed-time-path formalism of Zhou et al. (1980) [S4] provided yet another approach to non-equilibrium dynamics that would eventually connect to the stochastic quantization program.

Parisi and Wu’s contribution [S5] was to recognize that by adding a fictitious time dimension and using the action functional as a drift term, one could generate the Euclidean quantum field theory as the equilibrium limit of a well-defined stochastic process. The key distinction from Nelson’s stochastic mechanics is that the stochastic time $s$ is not physical time; it is an auxiliary parameter that serves as a computational device. The equilibrium limit $s \to \infty$ recovers the standard Euclidean theory, while finite-$s$ dynamics may carry physical information about approach to equilibrium and relaxation.

The formalism was rapidly developed in the 1980s. Seiler (1984) [S6] studied the gauge-fixing properties, Zinn-Justin (1986) [S7] established the renormalization theory, and Rumpf (1986) [S8] made the first attempt at stochastic quantization of Einstein gravity. The comprehensive review by Damgaard and Hüffel (1987) [S10]—at 170 pages, effectively a textbook in review-article form—remains the standard reference for the classical theory. Namiki’s monograph (1992) [S12] and accompanying overview (1993) [S14] provide a complementary treatment with emphasis on foundational issues.


2. Euclidean Stochastic Quantization: The Parisi–Wu Formalism

2.1 The Langevin equation and its conventions

The central object is a field $\Phi(s, x)$ that depends on spacetime coordinates $x \in \mathbb{R}^n$ and a fictitious stochastic time $s \in [0, \infty)$. Given a Euclidean action functional $S_E[\Phi]$, the Langevin equation reads

\[\frac{\partial \Phi(s,x)}{\partial s} = -\frac{\delta S_E[\Phi]}{\delta \Phi(s,x)} + \eta(s,x),\]

where $\eta(s,x)$ is Gaussian white noise satisfying

\[\langle \eta(s,x) \rangle = 0, \qquad \langle \eta(s,x)\, \eta(s',x') \rangle = 2\,\delta(s-s')\,\delta^{(n)}(x-x').\]

A note on sign conventions. The sign of the drift term is critical and is a source of some confusion in the literature. The convention adopted here—a minus sign in front of $\delta S_E / \delta \Phi$—is required for the stationary distribution to be proportional to $e^{-S_E}$. The drift pushes the field toward configurations that minimize the Euclidean action (i.e., toward classical solutions), while the noise drives fluctuations around them. In equilibrium, these effects balance to produce the path integral measure.

Some authors write the Langevin equation with a plus sign, $\partial_s \Phi = +\delta S/\delta \Phi + \eta$, which is consistent if $S$ is defined with the opposite overall sign from $S_E$ (e.g., as $S = -S_E$, so that the Boltzmann weight is $e^{+S}$), or if the equation is written in Minkowski signature where the action carries a factor of $i$. When reading the literature, one must check which convention is in force. In these notes, $S_E$ always denotes the Euclidean action with the standard sign, so that the classical equations of motion correspond to $\delta S_E/\delta \Phi = 0$ as an extremum condition, and the Langevin drift carries a minus sign.

2.2 The free scalar field as a warm-up

To fix ideas, consider a free massive scalar field in $n$ Euclidean dimensions with action

\[S_E[\Phi] = \frac{1}{2}\int d^n x \left[(\partial_\mu \Phi)^2 + m^2 \Phi^2\right].\]

The functional derivative is $\delta S_E / \delta \Phi = (-\nabla^2 + m^2)\Phi$, so the Langevin equation becomes

\[\frac{\partial \Phi(s,x)}{\partial s} = (\nabla^2 - m^2)\Phi(s,x) + \eta(s,x).\]

This is a linear SPDE—a stochastic heat equation with a mass term. Its solution can be written in Fourier space: for each momentum mode $\tilde{\Phi}(s,k)$,

\[\frac{\partial \tilde{\Phi}(s,k)}{\partial s} = -(k^2 + m^2)\tilde{\Phi}(s,k) + \tilde{\eta}(s,k).\]

This is an Ornstein–Uhlenbeck process with relaxation rate $\gamma_k = k^2 + m^2$. The two-point correlator at equal stochastic times, in the limit $s \to \infty$, is

\[\lim_{s\to\infty} \langle \tilde{\Phi}(s,k)\,\tilde{\Phi}(s,k')\rangle = \frac{1}{k^2 + m^2}\,\delta^{(n)}(k+k'),\]

which is precisely the Euclidean propagator. The stochastic process has found the correct quantum two-point function without any reference to path integrals or canonical commutation relations.

2.3 Interacting theories and the perturbative expansion

For an interacting theory with potential $U(\Phi)$—the standard example being $U(\Phi) = \lambda \Phi^4 / 4!$—the Langevin equation becomes nonlinear:

\[\frac{\partial \Phi}{\partial s} = (\nabla^2 - m^2)\Phi - U'(\Phi) + \eta.\]

The stochastic perturbation theory proceeds by treating $U’(\Phi)$ as a perturbation of the free Langevin dynamics. This generates a diagrammatic expansion (sometimes called the “stochastic diagram” expansion) that can be shown, order by order, to reproduce the standard Feynman diagram expansion for Euclidean correlators. The proof of this equivalence is most cleanly accomplished via the Fokker–Planck equation, to which we now turn.


3. Formal Equivalence: The Fokker–Planck Approach

3.1 The Fokker–Planck equation for field distributions

The Langevin equation defines a stochastic process on the space of field configurations. The probability distribution $P[\Phi, s]$ over field configurations at stochastic time $s$ satisfies the functional Fokker–Planck equation:

\[\frac{\partial P[\Phi, s]}{\partial s} = \int d^n x \, \frac{\delta}{\delta \Phi(x)} \left[\frac{\delta S_E}{\delta \Phi(x)} P[\Phi,s] + \frac{\delta P[\Phi,s]}{\delta \Phi(x)}\right].\]

This is a functional diffusion equation with drift. Its structure is that of a continuity equation in function space: the first term represents the deterministic drift toward the action minimum, and the second term represents diffusion due to the noise.

3.2 The stationary distribution

The crucial observation is that the distribution

\[P_{\mathrm{eq}}[\Phi] = \frac{1}{Z}\,e^{-S_E[\Phi]}, \qquad Z = \int \mathcal{D}\Phi\, e^{-S_E[\Phi]},\]

is a stationary solution of the Fokker–Planck equation. To verify this, substitute $P_{\mathrm{eq}}$ and note that

\[\frac{\delta P_{\mathrm{eq}}}{\delta \Phi(x)} = -\frac{\delta S_E}{\delta \Phi(x)}\,P_{\mathrm{eq}}[\Phi],\]

so the integrand of the Fokker–Planck equation vanishes pointwise. This is the detailed-balance condition: the drift and diffusion currents cancel in equilibrium.

Establishing that $P_{\mathrm{eq}}$ is the unique stationary distribution, and that the Langevin dynamics converges to it from generic initial conditions, requires more care. For the free theory, this follows from the explicit Ornstein–Uhlenbeck solution. For interacting theories, one can establish convergence perturbatively (each order in the coupling converges) or invoke general results on the ergodicity of infinite-dimensional diffusion processes. The mathematical literature on this point is not entirely settled for all theories of interest; see the discussion in Damgaard and Hüffel [S10], §5.

3.3 Correlators and the equilibrium limit

Once equilibrium is established, stochastic averages over the noise $\eta$ reproduce Euclidean path integral expectations:

\[\lim_{s\to\infty} \langle \Phi(s,x_1)\cdots\Phi(s,x_n)\rangle_\eta = \frac{1}{Z}\int \mathcal{D}\Phi\,\Phi(x_1)\cdots\Phi(x_n)\,e^{-S_E[\Phi]}.\]

The left-hand side is computed by solving the Langevin equation (numerically or perturbatively) and averaging over noise realizations. The right-hand side is the standard Euclidean correlator. This equality is the fundamental result of stochastic quantization.

3.4 Finite stochastic time and non-equilibrium dynamics

An important feature of the formalism is that the approach to equilibrium itself carries information. At finite stochastic time $s$, the distribution $P[\Phi, s]$ depends on the initial condition $P[\Phi, 0]$. The relaxation rates—the eigenvalues of the Fokker–Planck operator—are related to the spectrum of the theory. In lattice simulations, the rate of convergence to equilibrium (autocorrelation time) is a practical concern that directly impacts computational cost.

The finite-$s$ dynamics also provides a natural framework for studying non-equilibrium phenomena. This connection has been developed by several groups; the cluster of papers by Berges and collaborators [S15, S18] and Hirayama et al. [S16, S17] explores the interface between classical-statistical field theory and quantum dynamics, using stochastic methods to simulate real-time quantum field dynamics in regimes where classical-statistical approximations are valid.

A caveat on reference S15 (Berges, 2005): the paper archived as hep-lat/0508030 is specifically a contribution on simulating nonequilibrium quantum fields via stochastic quantization methods, demonstrating convergence properties and addressing numerical issues. It cites and builds on Berges’s broader pedagogical reviews of functional methods for non-equilibrium QFT, but should not be confused with those reviews themselves. The distinction matters: S15 is closer to a “numerical method demonstration with convergence analysis” than a “broad tutorial survey,” and its value for the present program lies in the practical lessons about lattice implementation.


4. Gauge Theories: Quantization Without Gauge Fixing

4.1 The Parisi–Wu observation

The original motivation for stochastic quantization was Parisi and Wu’s observation [S5] that gauge theories can be treated without explicit gauge fixing. Consider a gauge theory with action $S_E[A]$ that is invariant under gauge transformations $A_\mu \to A_\mu^g$. The Langevin equation

\[\frac{\partial A_\mu(s,x)}{\partial s} = -\frac{\delta S_E}{\delta A_\mu(s,x)} + \eta_\mu(s,x)\]

respects the gauge symmetry in the sense that the drift $-\delta S_E / \delta A_\mu$ is tangent to the gauge orbit (it vanishes for pure gauge directions), while the noise $\eta_\mu$ has components both along and transverse to the orbit. The longitudinal (gauge) components undergo a random walk, exploring the gauge orbit without bound, but this unbounded diffusion contributes nothing to gauge-invariant observables.

4.2 Longitudinal diffusion and gauge damping

The unbounded diffusion of gauge modes is, in practice, a nuisance: it causes the numerical integration to explore increasingly large gauge transformations, degrading the signal-to-noise ratio. Zwanziger (1981) introduced a gauge-damping term that confines the longitudinal diffusion without altering the gauge-invariant content of the theory. The modified Langevin equation takes the form

\[\frac{\partial A_\mu}{\partial s} = -\frac{\delta S_E}{\delta A_\mu} - \frac{\alpha}{2}\partial_\mu(\partial \cdot A) + \eta_\mu,\]

where $\alpha$ is a damping parameter. The additional term is a gradient in field space along the gauge orbit and therefore does not affect gauge-invariant correlators. It does, however, prevent the longitudinal modes from wandering to infinity, making numerical simulations feasible.

4.3 Gribov copies and the stochastic perspective

In the Faddeev–Popov approach, the Gribov ambiguity—the existence of multiple gauge-equivalent configurations satisfying the gauge condition—is a fundamental obstruction. In stochastic quantization, the issue takes a different form. The Langevin dynamics does not select a gauge; it averages over the entire gauge orbit. The question of whether this averaging is well-defined (does it produce the correct gauge-invariant correlators?) is answered affirmatively in perturbation theory but remains an open question nonperturbatively for theories where Gribov copies are abundant. Seiler’s analysis [S6] addresses some aspects of this question.


5. Notation and Conventions for Curved Backgrounds

Before turning to the numerical relativity connection, we establish the notation for fields on curved backgrounds. This notation is used throughout the remainder of the notes.

5.1 Spacetime and internal indices

We consider a field $\Phi^\mathfrak{j}(x)$ living on a manifold $(\mathcal{M}, g_{ab})$ and carrying internal indices $\mathfrak{j}, \mathfrak{k}, \ldots$ associated with a fiber bundle $E$ over $\mathcal{M}$. The spacetime coordinates are $x^a = (t, x^i)$ with $a = 0, 1, \ldots, n-1$ and $i = 1, \ldots, n-1$. The conventions are:

  • Spacetime metric: $g_{ab}(x)$, with determinant $g$ and inverse $g^{ab}$.
  • Internal metric: $\mathfrak{g}_{\mathfrak{j}\mathfrak{k}}$, governing traces over internal symmetry indices.
  • Gauge-covariant derivative: $\nabla_a = \partial_a - A_a$, where $A_a$ is a connection on $E$.

5.2 The general scalar field action

The paradigmatic action for a scalar field with internal structure on a curved background is

\[S[\Phi] = \int d^n x \, \sqrt{|g|} \left\{\mathfrak{g}_{\mathfrak{j}\mathfrak{k}} \, \nabla_a \Phi^\mathfrak{j} \, g^{ab} \, \nabla_b \Phi^\mathfrak{k} - U(\Phi)\right\}.\]

This action encodes: background spacetime geometry (through $g_{ab}$ and $\sqrt{\vert g \vert}$); gauge symmetry (through the connection $A_a$ in $\nabla_a$); internal symmetry structure (through $\mathfrak{g}_{\mathfrak{j}\mathfrak{k}}$); and self-interactions (through the potential $U(\Phi)$).

5.3 The curved-background Langevin equation

The stochastic quantization of this system proceeds by introducing fictitious time $s$ and writing

\[\frac{\partial \Phi^\mathfrak{j}(s,x)}{\partial s} = -\frac{1}{\sqrt{|g|}}\frac{\delta S_E[\Phi]}{\delta \Phi^\mathfrak{j}(x)} + \eta^\mathfrak{j}(s,x),\]

where $S_E$ is the Euclidean continuation of the action and the noise correlator is

\[\langle \eta^\mathfrak{j}(s,x)\, \eta^\mathfrak{k}(s',x') \rangle = 2\,\mathfrak{g}^{\mathfrak{j}\mathfrak{k}} \, \frac{\delta^{(n)}(x-x')}{\sqrt{|g|}}\,\delta(s-s').\]

The factor of $1/\sqrt{\vert g \vert}$ in the drift and noise ensures covariance under coordinate transformations on $\mathcal{M}$, so that the equilibrium distribution remains $P_{\mathrm{eq}}[\Phi] \propto e^{-S_E[\Phi]}$ with the correct functional measure.


6. Dimensional Hierarchy of Applications

One of the virtues of the general framework in §5 is that it specializes cleanly to theories in different dimensions. This section surveys the key targets, organized by spacetime dimension, that the formalism addresses.

6.1 $n = 1$: quantum mechanics

In the limit of zero spatial dimensions ($n = 1$), the “field” $\Phi$ reduces to a particle coordinate $x^\mathfrak{i}(\tau)$ depending on a single parameter $\tau$ (either Newtonian time or proper time along a worldline). The action becomes the mechanical action

\[S[x] = \int d\tau \left\{\mathfrak{g}_{\mathfrak{i}\mathfrak{j}} \, \dot{x}^\mathfrak{i} \dot{x}^\mathfrak{j} + A_\mathfrak{k}\,\dot{x}^\mathfrak{k} - U(x)\right\},\]

and the Langevin equation is

\[\frac{\partial x^\mathfrak{i}}{\partial s} = -\frac{\delta S}{\delta x^\mathfrak{i}} + \xi^\mathfrak{i}(s, \tau).\]

This is stochastic quantization of quantum mechanics: a particle diffusing in its configuration space, with the classical action providing the drift. The equilibrium distribution $e^{-S[x]}$ is the Euclidean path integral weight, and stochastic averages reproduce quantum mechanical matrix elements.

This $n = 1$ case is valuable both as a testing ground for numerical methods (one can compare against exact results) and as a pedagogical entry point. It connects to: the Feynman–Kac formula (the precise statement of the path-integral/stochastic-process correspondence in quantum mechanics); the Caldeira–Leggett model of quantum Brownian motion, studied extensively by Hu et al. [S11, S13] using influence-functional methods; and the foundations of non-equilibrium statistical mechanics.

6.2 $n = 2$: conformal field theory

In $1+1$ dimensions (or 2 Euclidean dimensions), the infinite-dimensional Virasoro symmetry constrains the dynamics severely. Stochastic quantization in this setting provides an alternative route to conformal field theory correlators that may offer computational advantages for theories where the conformal bootstrap is not easily applied. The connection to the theory of stochastic Loewner evolution (SLE), which describes random curves in two dimensions, is suggestive but not fully developed in the present notes.

6.3 $n = 3$: topological field theories and the fractional quantum Hall effect

In $2+1$ dimensions, the relevant theories include Chern–Simons gauge theory and its application to the fractional quantum Hall effect (FQHE). Stochastic quantization of Chern–Simons theory is interesting because the action is first-order (linear in time derivatives) and topological, presenting different challenges from the standard scalar or Yang–Mills cases. The topological nature of the theory means that the stochastic quantization must respect the topological invariance, adding a layer of subtlety.

6.4 $n = 4$: triviality, confinement, and the Standard Model

In $3+1$ dimensions, the primary targets are $\phi^4$ theory (where the “triviality” problem—the vanishing of the renormalized coupling in the continuum limit—is a fundamental issue in the foundations of quantum field theory) and non-Abelian gauge theories (where confinement, chiral symmetry breaking, and the full phenomenology of QCD are at stake). For $\phi^4$ triviality, stochastic quantization provides an independent numerical approach to studying the renormalization group flow and the continuum limit, complementing the extensive lattice Monte Carlo literature.

6.5 Fractional noise and anomalous diffusion

The standard Langevin equation uses Gaussian white noise, which corresponds to Markovian dynamics. A natural generalization replaces white noise with fractional Gaussian noise, producing anomalous diffusion in stochastic time. The physical interpretation and mathematical properties of this extension are not yet fully understood, but it opens connections to the theory of fractional calculus and anomalous transport that may have applications in disordered systems and non-equilibrium phenomena.


7. The Numerical Relativity Connection

This section develops what is perhaps the most distinctive aspect of these notes: the connection between stochastic quantization and the methods of numerical relativity. The basic idea is that the stochastic PDE defining the quantization of a field on a curved background can be solved using the same $3+1$ decomposition and numerical techniques that are standard in numerical relativity.

7.1 The $3+1$ decomposition for a scalar field

Consider a free scalar field $\Phi$ on a spherically symmetric spacetime. The standard $3+1$ (ADM) decomposition introduces auxiliary variables

\[\varphi \equiv \frac{\partial \Phi}{\partial r}, \qquad \Pi \equiv \frac{a}{\alpha}\frac{\partial \Phi}{\partial t},\]

where $\alpha$ is the lapse function and $a$ is the radial metric function (so that the radial part of the spatial metric is $ds^2_{\mathrm{rad}} = a^2 dr^2$). The evolution equations are

\[\frac{\partial \Phi}{\partial t} = \frac{\partial}{\partial r}\left[\frac{\alpha}{a}\,\Pi\right], \qquad \frac{\partial \Pi}{\partial t} = \frac{3}{r^2}\frac{\partial}{\partial r}\left[r^2\frac{\alpha}{a}\,\varphi\right],\]

and the lapse $\alpha$ and metric function $a$ satisfy the Einstein constraint equations. This system is the starting point for standard numerical relativity codes for spherically symmetric spacetimes.

7.2 Adding stochastic noise: QFT on curved backgrounds

To perform stochastic quantization of the scalar field on this background, one adds a noise term to the evolution in fictitious time $s$. Schematically, the $3+1$-decomposed field variables are promoted to functions of $(s, t, r)$, and the Langevin drift is constructed from the appropriate functional derivative of the Euclidean action on the curved background. The resulting system is a set of coupled SPDEs that can be integrated using methods familiar from numerical relativity: method of lines, finite differencing, and constraint-preserving boundary conditions.

The key point is that existing numerical relativity infrastructure—including adaptive mesh refinement, characteristic extraction, and constraint monitoring—can be repurposed for stochastic quantization calculations on curved backgrounds. This is a significant practical advantage.

7.3 Fixed backgrounds: Schwarzschild and de Sitter

When the background geometry is fixed (i.e., the metric is a known solution of the Einstein equations, not evolved dynamically), the stochastic quantization of a scalar field on that background reduces to solving a linear (or nonlinear, if there are self-interactions) SPDE on a given manifold.

Schwarzschild background. Fixing $\alpha$ and $a$ to their Schwarzschild values provides a laboratory for studying quantum fields in black hole spacetimes: the Hawking effect, the Unruh effect, vacuum fluctuations near horizons, and the information paradox. Iso and Okazawa (2011) [S25] have studied stochastic equations on black hole backgrounds in this spirit, and de Aguiar et al. (2008) [S19] studied scalar fields on de Sitter backgrounds.

De Sitter background. The de Sitter case is directly relevant to inflationary cosmology: quantum fluctuations of scalar fields during inflation seed the primordial density perturbations that grow into large-scale structure. Stochastic quantization provides a natural framework for computing these fluctuations, complementing the standard Bunch–Davies vacuum approach.

7.4 Toward dynamical backgrounds: the three-tiered conceptual structure

When one contemplates making the background geometry dynamical—solving the Einstein equations simultaneously with the stochastic matter field equations—one encounters a crucial conceptual distinction that the literature does not always make clear. There are (at least) three distinct programs that involve “stochastic” and “gravity,” and conflating them leads to confusion:

Tier 1: Stochastic quantization of matter on a fixed curved background. This is the program described in §7.2–7.3. The geometry is classical and given; the matter field is quantized via the Langevin equation. The output is QFT correlators on curved spacetime. This is well-defined and computationally feasible with existing methods.

Tier 2: Stochastic gravity (the Einstein–Langevin equation). This is the semiclassical program developed by Hu and Verdaguer [S20, S30] and reviewed in their 2008 Living Reviews in Relativity article and 2020 Cambridge monograph. Here, the metric fluctuations are driven by quantum stress-energy noise: the expectation value of the stress-energy tensor sources the semiclassical Einstein equation, and its fluctuations (the noise kernel) source metric perturbations via an Einstein–Langevin equation. This is an effective field theory approach to backreaction; it does not quantize the metric in the sense of Tier 3. The stochastic element is the quantum noise of matter, not a fictitious Langevin time for the metric. Applications include the stability of Minkowski space under quantum corrections, structure formation in cosmology, and black hole backreaction including Hawking radiation.

The related work by Antonio dos Reis et al. (2018) [S29] on semiclassical gravity connects to this tier, providing further development of the formalism for practical calculations.

Tier 3: Stochastic quantization of gravity itself. This is the most ambitious program: applying the Parisi–Wu procedure to the Einstein–Hilbert action (or some suitable extension), so that the metric $g_{ab}$ itself undergoes Langevin dynamics in fictitious time and the equilibrium distribution is $e^{-S_{\mathrm{EH}}[g]}$. Rumpf (1986) [S8] made the first attempt in this direction. The program faces severe difficulties: the Einstein–Hilbert action is unbounded below (the conformal factor problem), the theory is perturbatively non-renormalizable, and the functional measure over metrics is not well-defined. Nevertheless, it remains a conceptually attractive target.

Why the distinction matters. The statement that stochastic quantization “extends naturally to gravity via the Einstein–Langevin equation” (which appears in some formulations of the program) is directionally suggestive but conflates Tier 2 and Tier 3. The Einstein–Langevin equation of stochastic gravity is not a Parisi–Wu Langevin equation for the metric; it is a semiclassical equation driven by matter noise. A fully stochastic quantization of gravity would require Tier 3, which is a far more speculative enterprise. The present notes aim to connect Tier 1 (matter on fixed backgrounds, using numerical relativity infrastructure) with the aspiration toward Tiers 2 and 3, while being explicit about what is established, what is plausible, and what is speculative.

7.5 The BTZ black hole as an intermediate target

The BTZ solution is exactly solvable classically (it is locally anti-de Sitter), exhibits genuine gravitational dynamics (horizons, thermodynamics, quasinormal modes), and lives in $2+1$ dimensions where the quantum theory is substantially simpler than in $3+1$. A program of stochastic quantization of matter coupled to BTZ gravity would test the formalism in a setting where exact results are available for comparison.


8. Complex Langevin Dynamics and the Sign Problem

8.1 The sign problem

The formalism described in §2–3 assumes a real, positive Euclidean action $S_E$, so that $e^{-S_E}$ is a well-defined probability weight. In many physically important cases, this assumption fails:

  • Finite-density QCD. At nonzero baryon chemical potential $\mu$, the fermion determinant becomes complex, and the Euclidean action acquires an imaginary part. Standard Monte Carlo methods, which rely on importance sampling with $e^{-S_E}$ as a probability measure, fail because $e^{-S_E}$ is no longer positive definite.
  • Real-time (Minkowski) field theory. The Minkowski action $S_M$ carries a factor of $i$ in the path integral weight $e^{iS_M}$, which oscillates rather than being exponentially damped. Direct simulation is impossible.
  • Theories with topological terms. The $\theta$-term in QCD, $\theta\,\mathrm{tr}(F \tilde{F})$, is purely imaginary in the Euclidean formulation.

These are instances of the “sign problem,” which is not merely a technical inconvenience but a fundamental obstruction to Monte Carlo simulation. It is known to be NP-hard in general.

8.2 Complex Langevin dynamics

The idea, going back to Parisi (1983) and Klauder (1983), is to complexify the field variables: $\Phi \to \Phi_R + i\,\Phi_I$, and extend the Langevin equation to the complexified field space. The drift $-\delta S/\delta \Phi$ is now complex (since $S$ is complex), so the real and imaginary parts of the field evolve according to coupled equations driven by the (real) noise:

\[\frac{\partial \Phi_R}{\partial s} = -\mathrm{Re}\,\frac{\delta S}{\delta \Phi}\bigg|_{\Phi_R + i\Phi_I} + \eta, \qquad \frac{\partial \Phi_I}{\partial s} = -\mathrm{Im}\,\frac{\delta S}{\delta \Phi}\bigg|_{\Phi_R + i\Phi_I}.\]

The hope is that stochastic averages computed from this complexified dynamics converge to the correct (complex-weighted) path integral expectations. Unlike the real case, this is not guaranteed.

8.3 Convergence and correctness criteria

The central difficulty with complex Langevin dynamics is that it can converge to wrong results. The conditions under which the method gives correct answers have been the subject of extensive investigation:

  • Aarts et al. (2009) [S24] identified specific instabilities in complex Langevin simulations and characterized the mechanisms by which the method can fail.
  • Guralnik and Pehlevan (2007) [S21] studied the relationship between complex Langevin dynamics and Schwinger–Dyson equations, providing criteria for when the stochastic averages satisfy the correct Ward identities.
  • Seiler (2017) [S27] provided a comprehensive analysis of the conditions for correctness, including the role of the probability distribution of the complexified field in the imaginary direction.
  • Berger et al. (2019) [S28] reviewed the sign problem and the status of complex Langevin as one approach among several (including Lefschetz thimble methods and density of states approaches).
  • Cai et al. (2021) [S31] proposed novel regularization techniques for complex Langevin dynamics, addressing some of the known convergence issues.

The current consensus is roughly this: complex Langevin works reliably when the probability distribution of the imaginary part of the complexified field is sufficiently well-localized (does not develop long tails into the imaginary direction). When this condition is satisfied—which can sometimes be enforced by gauge cooling or adaptive regularization—the method provides correct results. When it is violated, the method can produce systematically wrong answers with no internal diagnostic.

For the program described in these notes, the complex Langevin extension is relevant in two contexts: the simulation of real-time (Minkowski) field dynamics on curved backgrounds, and the treatment of the conformal-factor problem in the stochastic quantization of gravity (Tier 3 of §7.4), where the Euclidean Einstein–Hilbert action is not bounded below.


9. Stochastic Gravity: The Einstein–Langevin Equation

This section develops the stochastic gravity program of Hu and Verdaguer in more detail, since it represents the most mature gravity-facing application of stochastic methods and is often conflated with the Parisi–Wu formalism.

9.1 Semiclassical gravity and its limitations

Semiclassical gravity replaces the classical stress-energy tensor on the right-hand side of the Einstein equations with its quantum expectation value:

\[G_{ab} = 8\pi G \, \langle \hat{T}_{ab} \rangle_{\mathrm{ren}}.\]

This is the mean-field approximation: the geometry responds to the average quantum matter, ignoring fluctuations. It is adequate when quantum fluctuations of the stress-energy are small compared to the mean, but it breaks down in precisely the regimes of greatest interest: near black hole horizons, during the late stages of Hawking evaporation, and in the early universe.

9.2 The noise kernel and the Einstein–Langevin equation

Stochastic gravity improves on semiclassical gravity by incorporating the leading quantum fluctuations of the stress-energy tensor. The key object is the noise kernel, defined as the symmetrized two-point function of the stress-energy fluctuations:

\[N_{abcd}(x, x') = \frac{1}{2}\langle \{\hat{t}_{ab}(x), \hat{t}_{cd}(x')\}\rangle, \qquad \hat{t}_{ab} \equiv \hat{T}_{ab} - \langle \hat{T}_{ab}\rangle.\]

The Einstein–Langevin equation is

\[G_{ab} = 8\pi G \left[\langle \hat{T}_{ab}\rangle_{\mathrm{ren}} + \xi_{ab}\right],\]

where $\xi_{ab}$ is a stochastic tensor whose correlator is the noise kernel: $\langle \xi_{ab}(x)\,\xi_{cd}(x’)\rangle = N_{abcd}(x,x’)$. The metric perturbations induced by $\xi_{ab}$ represent the leading-order effect of quantum stress-energy fluctuations on the geometry.

9.3 Applications

The stochastic gravity program has been applied to several problems [S20, S30]:

The stability of flat spacetime under quantum corrections was one of the first applications: does Minkowski space remain stable when quantum matter fluctuations are included? The analysis shows that it does, with the metric perturbations being small and well-behaved.

In cosmology, the noise kernel sources metric perturbations during inflation that contribute to the primordial power spectrum. This provides a stochastic-gravity derivation of cosmological perturbation theory that complements the standard approach based on quantum fluctuations of the inflaton.

For black holes, the noise kernel drives metric fluctuations near the horizon that are relevant to the information paradox and the endpoint of Hawking evaporation. The stochastic gravity approach provides a systematic framework for studying backreaction beyond the semiclassical approximation.

9.4 Relationship to Parisi–Wu

It is worth restating the relationship explicitly. In Parisi–Wu stochastic quantization, one introduces a fictitious time $s$ and noise $\eta(s,x)$ as a computational device for generating the path integral measure; the noise has no physical interpretation, and the limit $s \to \infty$ is taken to recover the standard quantum theory. In stochastic gravity, the noise $\xi_{ab}$ represents physical quantum fluctuations of the stress-energy tensor; there is no fictitious time, and no $s \to \infty$ limit. The two frameworks are related—both involve stochastic differential equations driven by Gaussian noise, and both connect classical stochastic dynamics to quantum physics—but they are not the same thing. A coherent research program can draw on both, but must not conflate them.


Several adjacent developments in the literature connect to and illuminate the stochastic quantization program.

10.1 The Keldysh formalism

The closed-time-path (Schwinger–Keldysh) formalism [S4, S22] provides a path integral framework for real-time, non-equilibrium quantum field theory. Kamenev and Levchenko (2009) [S22] give a modern treatment connecting the Keldysh formalism to $\sigma$-models and disorder physics. The relationship to stochastic quantization is through the MSR formalism [S1]: the Janssen–De Dominicis representation of classical stochastic dynamics is structurally similar to the Keldysh path integral, with the stochastic noise playing the role of the “quantum” or “ghost” field. De Dominicis and Giardina (2009) [S23] develop this connection in the context of random fields and spin glasses.

10.2 Influence functionals and quantum Brownian motion

The work of Hu, Paz, and Zhang on quantum Brownian motion [S11, S13] developed the Feynman–Vernon influence functional for studying the dynamics of a quantum system coupled to a thermal bath. The influence functional encodes the effect of the bath on the system as a non-local (in time) action that generates both dissipation and noise. This is directly relevant to stochastic quantization: the thermal bath plays the role of the noise source, and the effective dynamics of the system is described by a quantum Langevin equation. The Caldeira–Leggett model is the paradigmatic example, and its analysis provides insights into decoherence, thermalization, and the quantum-to-classical transition that inform the interpretation of stochastic quantization.

10.3 Stochastic regularization and the Langevin equation

One motivation for stochastic quantization that is sometimes overlooked is its potential as a regularization scheme. The noise correlation length in stochastic time $s$ provides a natural ultraviolet regulator. Goldschmidt (1987) [S9] explored this idea in the context of magnetization decay, and the Damgaard–Hüffel review [S10] includes a detailed discussion of stochastic regularization as an alternative to dimensional regularization or lattice regularization.

10.4 Landau–Ginzburg theory and classical-statistical methods

Cassol-Seewald et al. (2012) [S26] applied stochastic methods to Landau–Ginzburg theory, connecting the stochastic quantization framework to the theory of phase transitions and critical phenomena. This work illustrates the broader point that stochastic quantization is not merely a formal trick for generating path integrals, but a genuine extension of the toolkit of statistical field theory.


11. Open Problems

11.1 Mathematical status of the equivalence

The equivalence between stochastic quantization and Euclidean path integral quantization is established at the following levels:

Perturbative equivalence: The stochastic diagram expansion reproduces the standard Feynman diagram expansion order by order in the coupling. This is proven for scalar theories, QED, and Yang–Mills theory in the Damgaard–Hüffel review [S10].

Fokker–Planck equilibrium: The distribution $e^{-S_E}$ is a stationary solution of the functional Fokker–Planck equation. Uniqueness of this stationary distribution is established under conditions that are satisfied for free and weakly coupled theories.

Non-perturbative status: For strongly coupled theories, the rigorous proof of equivalence remains incomplete. The lattice regularization provides a well-defined finite-dimensional system for which ergodicity theorems can be applied, but the continuum limit is a separate (and harder) question.

11.2 Computational considerations

For numerical implementation:

  1. The discretization of stochastic time introduces systematic errors (analogous to the step-size dependence in any numerical ODE/SDE integration). The choice of discretization scheme (Itô, Stratonovich, or other) affects the convergence properties and must be matched to the continuum convention.

  2. Autocorrelation times can be severe, especially near critical points where the correlation length diverges (critical slowing down). Stochastic quantization shares this problem with Monte Carlo methods; Fourier acceleration and multigrid techniques have been explored as remedies.

  3. For gauge theories, the gauge-damping parameter $\alpha$ must be chosen carefully: too small, and the longitudinal modes wander; too large, and the dynamics is dominated by the gauge-fixing term rather than the physical action.

12. Prospects for Computational Projects

The notes outlined here suggest a staged computational program:

Stage 1: Matter fields on fixed backgrounds. Implement the Langevin equation for scalar (and eventually spinor and gauge) fields on Schwarzschild and de Sitter backgrounds using $3+1$ numerical relativity infrastructure. Compare against known results (e.g., Hawking temperature, de Sitter particle production rates). This stage is well-defined and achievable with existing tools.

Stage 2: Stochastic gravity calculations. Implement the Einstein–Langevin equation for metric perturbations driven by quantum stress-energy noise on fixed backgrounds. Compute the noise kernel, solve for the induced metric fluctuations, and compare against the Hu–Verdaguer results [S20, S30]. This requires additional infrastructure (the noise kernel is a bi-tensor that must be computed from the quantum field theory on the background) but is conceptually straightforward.

Stage 3: BTZ as a testing ground. Implement stochastic quantization of a scalar field coupled to $2+1$ gravity in the BTZ sector. This tests the formalism in a controlled setting where exact classical results are available and the quantum theory is better understood than in $3+1$ dimensions.

Stage 4: Full $3+1$ gravity. Extending stages 1-3 to the stochastic quantization of the Einstein–Hilbert action requires addressing: A. The conformal factor problem, B. Non-renormalizability, and C. Definition of the functional measure. This long-term target requires new ideas beyond the current framework posing a highly speculative and ambitious committment.


Bibliography

Classical stochastic dynamics and non-equilibrium methods

IDLinkNotes
S1Martin et al “Statistical Dynamics of Classical Systems” 1973Mathematical foundation for stochastic quantization.
S2Hohenberg and Halperin “Theory of Dynamic Critical Phenomena” 1977Provides the technology of Langevin equations coupled to order parameter fields.
S3Fox “Gaussian stochastic processes in physics” 1978Mathematical foundations of the noise terms in stochastic quantization.
S4Zhou et al “Closed time path Green’s functions and critical dynamics” 1980Closed time path (CTP) Green’s function formalism. Connects to Schwinger–Keldysh formalism

The Parisi–Wu formalism and its development

IDLinkNotes
S5Parisi Wu “Perturbation Theory Without Gauge Fixing” 1980Proposes a stochastic quantization approach that avoids explicit gauge fixing by introducing auxiliary time variables.
S6Seiler “Stochastic Quantization and Gauge Fixing” 1984Analyzes the relationship between stochastic quantization techniques and conventional gauge-fixing procedures.
S7J. Zinn-Justin “Renormalization and stochastic quantization” 1986Investigates the renormalization of quantum field theories within the stochastic quantization framework.
S8Rumpf “Stochastic Quantization of Einstein Gravity” 1986Applies stochastic quantization to the gravitational field, exploring potential routes to quantize Einstein’s theory.

Reviews and monographs

IDLinkNotes
S10Damgaard and Huffel “Stochastic Quantization” 1987Provides a comprehensive review of stochastic quantization techniques and their applications in field theory.
S12Namiki “Stochastic Quantization” 1992Monograph detailing the foundations and various applications of stochastic quantization in quantum field theory.
S14Namiki “Basic Ideas” 1993Outlines the fundamental concepts underlying stochastic quantization, with emphasis on gauge theories.

Quantum Brownian motion and open quantum systems

IDLinkNotes
S9Goldschmidt “Decay of Magnetization” 1987Studies the time-dependent decay of magnetization near critical points using stochastic and renormalization methods.
S11Hu et al “Quantum Brownian Motion I” 1992Develops an influence functional approach for analyzing quantum Brownian motion of a system coupled to a thermal bath.
S13Hu et al “Quantum Brownian Motion II” 1993Extends the quantum Brownian motion framework to relativistic fields and investigates decoherence processes.

Non-equilibrium QFT and classical-statistical simulations

IDLinkNotes
S15Berges “Noneqilbrium QFT” 2005Reviews theoretical techniques for studying nonequilibrium quantum field dynamics using functional methods.
S16Hirayama “Classical Simulation of QFT I” 2005Proposes a classical lattice simulation scheme for out-of-equilibrium quantum fields based on classical statistical analogies.
S17Hirayama et al “Classical Simulation of QFT II” 2005Extends classical simulation methods to non-Abelian gauge theories, demonstrating feasibility in lattice frameworks.
S18Berges and Sexty “Gauge theory simulations” 2007Presents real-time lattice simulations of gauge theories using classical-statistical approximations to explore thermalization.
S32Hiryama “Classical simulation of QFT” 2024Reviews classical-statistical lattice methods for simulating real-time quantum field dynamics in various contexts.

Stochastic gravity

IDLinkNotes
S19de Aguiar et al “Scalar fields in de Sitter” 2008Examines stochastic dynamics and correlation functions of scalar fields propagating in de Sitter spacetime.
S20Hu and Verdaguer “Stochastic Gravity” 2008Provides a comprehensive review of stochastic gravity, where metric fluctuations are driven by quantum stress-energy noise.
S25Iso and Okazawa “Stochastic equations on blackhole backgrounds” 2011Formulates stochastic differential equations for quantum fields in black hole spacetimes to explore horizon fluctuations.
S29Antonio dos Reis et al “Semiclassical Gravity” 2018Develops semiclassical approximation schemes for gravity coupled to quantum fields, focusing on backreaction effects.
S30Hu and Verdaguer “Semiclassical and Stochastic Gravity” 2020Updates the theoretical framework blending semiclassical gravity with stochastic metric fluctuations for early-universe studies.

Complex Langevin and the sign problem

IDLinkNotes
S21Guralnik and Pehlevan “Complex Langevin and Shwinger Dyson” 2007Investigates complex Langevin methods as a means to solve nonperturbative Schwinger-Dyson equations in quantum field theory.
S22Kamenev and Levchenko “Keldysh and sigma models” 2009Reviews the Keldysh nonequilibrium formalism and its connection to nonlinear sigma models in condensed matter and field theory.
S23De Dominicis and Giardina “Random Fields and Spin Glasses” 2009Textbook covering theoretical methods—such as replica and functional approaches—for random fields and spin glass phenomena.
S24Aarts et al “Instabilities in complex Langevin”* 2009Studies numerical instabilities arising in complex Langevin simulations of quantum field theories with severe sign problems.
S26Cassol-Seewald et al “Landau Ginzburg” 2012Applies stochastic quantization techniques to the Landau–Ginzburg model and analyzes critical scaling behavior.
S27Seiler “Complex Langevin” 2017Reviews advances and challenges in complex Langevin methods for tackling complex action problems in lattice field theories.
S28Berger et al “Sign problem” 2019Surveys state-of-the-art approaches—such as Lefschetz thimbles and complex Langevin—to mitigate the sign problem in lattice QCD.
S31Cai et al “Regulariztion of Complex Langevin” 2021Proposes novel regularization techniques to improve convergence and reliability of complex Langevin simulations in field theories.
This post is licensed under CC BY 4.0 by the author.