Post

Quantum Foundations

Quantum Foundations

Notes on Quantum Foundations (Draft)

Abstract

Organized by interpretive tradition, these notes survey the foundations of quantum mechanics through a curated collection of 37 primary and contextual sources. The collection spans wave mechanics and pilot-wave theory, the Copenhagen formalism, statistical and information-theoretic reconstructions, Bell’s theorem and nonlocality, and the Everettian relative-state interpretation. For each source, the notes provide technical context, identify the interpretive work the source performs, and trace the conceptual threads connecting entries within and across sections. The notes are idiosyncratic rather than exhaustive. They reflect a particular path through the literature, chosen to illuminate the mathematical and physical structures underlying the interpretive debates. Where the collection has gaps — collapse theories, QBism, decoherent histories — these are documented and briefly discussed. The notes conclude with suggested reading paths for different audiences and a consolidated reference list.

Introduction

Quantum foundations is a subject in which the technical and the philosophical are unusually entangled. The formalism of quantum mechanics — Hilbert spaces, self-adjoint operators, the Born rule — has been settled in its mathematical essentials since the early 1930s. What remains unsettled, nearly a century later, is what the formalism means: what it says about the physical world, what constitutes a measurement, and whether the quantum state is a real thing or a bookkeeping device. These are not idle questions. They bear on quantum information theory and computing and the conceptual coherence of physics itself.

The present notes survey these questions through primary sources. The approach is deliberately source-first: rather than beginning with a synthetic overview and then pointing the reader toward the literature, these notes begin with the literature and build understanding through engagement with what the authors actually wrote. The organizing principle is interpretive tradition — wave mechanics, Copenhagen, statistical/information-theoretic, nonlocality, and Everettian — though this taxonomy is pedagogical rather than rigid. Several sources (Peres, for instance) resist neat classification.

The collection comprises 37 entries, organized into five sections. It is a personal sampling, not an exhaustive survey. The selection emphasizes foundational primary sources interspersed with historical and pedagogical texts that contextualize the technical arguments. The goal is to equip the reader with enough orientation to the primary literature that they can form their own judgments about the interpretive landscape.

A word on the gaps. The collection does not include sources on spontaneous collapse theories (GRW, CSL), QBism, relational quantum mechanics, consistent or decoherent histories, or several other post-1950s interpretive programs. These omissions are discussed in the final section, along with suggestions for how a reader might extend the collection.

A remark on the relevance of quantum foundations to scientific computing. The interpretive questions surveyed here are not merely philosophical: they have direct computational consequences. The pilot-wave interpretation implies a particular class of stochastic differential equations. Phase-space methods (Moyal, Wigner) underwrite semiclassical simulation techniques. The information-theoretic reconstruction program connects to quantum computing and quantum error correction through the structure of generalized probabilistic theories. Bell inequality violations are now routinely simulated as benchmarks for quantum hardware. And the measurement problem, far from being a philosophical curiosity, determines the computational cost of simulating quantum systems on classical hardware — the exponential overhead of exact simulation is a direct consequence of the superposition principle and its entanglement-generating dynamics.

These notes therefore serve a dual purpose within the present volume: They provide the conceptual and historical context for understanding why quantum computation is structured as it is, and they map the primary literature that a practitioner working at the interface of physics and computation should be familiar with.

Throughout, the notes distinguish between claims that are directly supported by the sources listed, claims that require reading the linked primary material, and claims that represent secondary or historical interpretation.

Wave Mechanics

This section traces the development of wave mechanics from de Broglie’s matter-wave hypothesis through Schrödinger’s equation to modern debates about the ontology of the quantum state. It is broader than the label “wave mechanics” suggests: it is also where the collection places pilot-wave realism (Bohm, de Broglie) and the $\psi$-ontology question (PBR theorem). The section contains 13 sources spanning 1925 to 2017.

The Matter-Wave Hypothesis

S1: de Broglie, “On the Theory of Quanta” (1925). De Broglie’s doctoral thesis introduces the concept of matter waves by extending Einstein’s light-quantum hypothesis to material particles: a particle of momentum $p$ is associated with a wave of wavelength $\lambda = p/h$. The thesis launches wave mechanics as a research program and provides the conceptual seed for both Schrödinger’s wave equation and the later pilot-wave interpretation. The key physical insight is the assertion that wave-particle duality is universal, not restricted to photons.

S2: Schrödinger, Collected Papers on Wave Mechanics (1928). This AMS collection compiles Schrödinger’s foundational papers developing the time-dependent and time-independent Schrödinger equations. The papers establish the mathematical formalism of wave mechanics, demonstrate its equivalence to Heisenberg’s matrix mechanics, and apply the theory to the hydrogen atom and other systems. The collection remains the primary source for understanding how wave mechanics was constructed, as opposed to how it is typically presented in textbooks (which tend to postulate the equation rather than derive it).

Bohm’s Pilot-Wave Interpretation

S3–S4: Bohm, “A Suggested Interpretation of the Quantum Theory in Terms of ‘Hidden’ Variables,” Parts I and II (1952; received 1951). Bohm’s two-part paper revives and extends de Broglie’s pilot-wave concept into a fully deterministic hidden-variable interpretation. In this framework, particles have definite positions at all times, guided by the wave function through a guidance equation. The wave function itself evolves via the Schrödinger equation; what is added is a specification of actual particle trajectories.

The key technical elements are the guidance equation $\dot{x} = \nabla S / m$ (where $S$ is the phase of the wave function written in polar form $\psi = R e^{iS/\hbar}$) and the quantum potential $Q = -\hbar^2 \nabla^2 R / (2mR)$, which mediates the nonlocal influence of the wave function on particle motion. The quantum potential is responsible for the nonclassical behavior of particles in this theory; it does not diminish with distance, which is why Bohm’s theory is explicitly nonlocal.

A point of historical and conceptual importance: Bohm explicitly argues that von Neumann’s celebrated impossibility proof for hidden variables does not apply to his scheme, because von Neumann’s proof assumes that the hidden variables must satisfy certain linearity conditions that Bohm’s theory violates. This observation, later sharpened by Bell, is one of the threads connecting the wave mechanics section to the Bell’s theorem section.

From a computational perspective, Bohm’s theory is significant because it transforms the quantum mechanical problem into a pair of coupled partial differential equations: the continuity equation for $R^2$ (the probability density) and a modified Hamilton–Jacobi equation for $S$ (the phase). These equations can be solved numerically, and Bohmian trajectory methods have found applications in quantum chemistry and quantum molecular dynamics, where they offer advantages for problems involving tunneling and interference. The computational cost of Bohmian methods scales differently from wavefunction-based methods, particularly in high-dimensional systems where the guidance equation provides a natural way to concentrate computational effort in regions of configuration space where the particle density is significant.

Schrödinger on Quantum Jumps

S5–S6: Schrödinger, “Are There Quantum Jumps?” Parts I and II (1952). These late papers represent Schrödinger’s mature critique of the discontinuous “jumps” associated with quantum measurement. Schrödinger remained dissatisfied throughout his career with the collapse postulate and the apparent discontinuity it introduces into an otherwise continuous dynamical theory. The papers are valuable less for technical novelty than for articulating — with characteristic clarity — the tension between the continuous evolution of the wave function and the apparently discontinuous outcomes of measurement. They belong to a tradition of measurement-skepticism that runs from Schrödinger through Everett to the decoherence program.

Technical and Historical Studies

S7: Wallstrom, “Inequivalence between the Schrödinger Equation and the Madelung Hydrodynamic Equations” (1994). Wallstrom’s paper addresses a technical subtlety in the relationship between Schrödinger’s equation and its hydrodynamic reformulation due to Madelung. The Madelung equations recast quantum mechanics as a kind of fluid dynamics, but Wallstrom shows that recovering the Schrödinger equation from the Madelung equations requires an additional quantization condition — specifically, a single-valuedness constraint on the phase that is not derivable from the hydrodynamic formulation alone. This result is important for pilot-wave and stochastic mechanics programs that attempt to derive quantum mechanics from more “classical” starting points: it demonstrates that the wave-mechanical structure is not as easily recovered from alternative formulations as one might hope.

S8: Bacciagaluppi and Valentini, Quantum Theory at the Crossroads (2006/2009; arXiv:quant-ph/0609184). This monograph is a large-scale historical and analytical reconstruction of the debates at the 1927 Solvay Conference. At 553 pages, it is substantially more than a historical narrative: the authors provide detailed technical analysis of the positions advocated by de Broglie, Born, Heisenberg, Schrödinger, and others, and trace how the pilot-wave interpretation was sidelined in favor of the Copenhagen consensus. The book functions both as intellectual history and as a rehabilitation of de Broglie’s pilot-wave program, documenting how the early debates contained technical resources — particularly in de Broglie’s work — that were not fully appreciated until Bohm’s revival in 1952 and the later development of Bohmian mechanics.

S9: Wheeler, “Schrödinger’s Argument” (2006). N. A. Wheeler’s pedagogical essay traces Schrödinger’s derivation of wave mechanics through the optico-mechanical analogy: the observation, going back to Hamilton in the 1830s, that geometric optics (ray theory) relates to wave optics as Hamilton–Jacobi theory relates to particle mechanics. The eikonal equation in optics, $\nabla F \cdot \nabla F = n^2$, is structurally identical to the Hamilton–Jacobi equation; Schrödinger completed the analogy by constructing wave mechanics as physical optics stands to geometric optics.

Wheeler extends this to Feynman’s path-integral formulation, showing how the two-point action function $S(x_1, t_1; x_0, t_0)$ generates the quantum propagator. The essay is pedagogically valuable for making explicit the chain of reasoning that led Schrödinger to his equation — reasoning that is almost entirely suppressed in standard textbook presentations.

S10: Joas and Lehner, “The Classical Roots of Wave Mechanics: Schrödinger’s Transformations of the Optical-Mechanical Analogy” (2009). This history-of-science article provides detailed analysis of how the analogy between geometric optics and classical mechanics served as a conceptual precursor to wave mechanics. It supplements the Wheeler essay by situating Schrödinger’s reasoning within the broader intellectual context of early 20th-century theoretical physics.

S11: Pusey, Barrett, and Rudolph (PBR), “On the Reality of the Quantum State” (2012; arXiv:1111.3328; Nature Physics). The PBR theorem is a no-go result demonstrating that the quantum state cannot be interpreted as merely epistemic — as representing only an observer’s knowledge or information — under a natural assumption called preparation independence. In the terminology introduced by Harrigan and Spekkens, PBR rules out $\psi$-epistemic models (where distinct quantum states can correspond to overlapping probability distributions over ontic states) in favor of $\psi$-ontic models (where the quantum state is a direct property of reality).

The theorem’s force depends critically on the preparation independence assumption: that independent preparations of systems yield independent ontic states. The philosophical import of the result is debated — advocates of epistemic interpretations have questioned whether preparation independence is physically well-motivated — but it has substantially shaped the post-2010 landscape of quantum foundations by sharpening the constraints on viable interpretive frameworks.

Placing PBR under “Wave Mechanics,” is not historically standard. However, the theorem bears directly on wavefunction realism and thereby on the ontological status of the wave-mechanical formalism.

S12: Lopes Coelho and Stachel (2013). This article in the European Journal of Physics examines historical influences on Schrödinger’s formulation, particularly from Hertz’s mechanics and Van Vleck’s work on classical–quantum correspondences. It provides useful contextual detail for understanding the intellectual milieu in which wave mechanics emerged.

S13: Landsman, “The Measurement Problem” (2017). This Springer chapter provides a modern treatment of the measurement problem, tracing its development from von Neumann’s formulation through decoherence theory to contemporary debates. It serves as a “modern synthesis” anchor for the wave mechanics section, connecting the historical sources to current technical work.

Summary

The 13 sources in this section provide a path from the origins of wave mechanics through pilot-wave theory to modern $\psi$-ontology debates. The emphasis on the de Broglie–Bohm line is not the standard wave-mechanical formalism (Schrödinger’s collected papers are included but not supplemented by, say, Dirac’s Principles). The Bacciagaluppi–Valentini monograph provides the kind of deep historical-technical analysis that rewards extended study and illustrates why de Broglie–Bohm has been included in consideration of wave mechanics. Further, inclusion of Wallstrom and PBR gives the view here a foundational character beyond “wave mechanics” as conventionally understood. Finally, the Wheeler essay and the Joas–Lehner article together illuminate the optico-mechanical analogy in a way that is not available in standard textbook treatments.

The Copenhagen Interpretation

This section covers the mathematical and interpretive framework developed by Bohr, Heisenberg, Dirac, and von Neumann, along with the group-theoretic foundations contributed by Weyl and Wigner. It contains nine sources spanning 1925 to 2009, mixing primary technical sources with biographical and historical works.

Dirac’s Contributions

S14: Dirac, “The Fundamental Equations of Quantum Mechanics” (1925). This early paper in the Proceedings of the Royal Society lays groundwork for the algebraic structure of quantum mechanics, establishing the quantization procedure that relates Poisson brackets to commutators. It marks the beginning of Dirac’s program to formulate quantum mechanics in terms of abstract algebraic relations rather than specific representations.

S15: Dirac, “On the Theory of Quantum Mechanics” (1926). Dirac introduces his transformation theory, providing a unified framework encompassing both Schrödinger’s wave mechanics and Heisenberg’s matrix mechanics. The paper develops the operator formalism and establishes the bra-ket notation’s conceptual precursors. Transformation theory was the first demonstration that the apparently different formulations of quantum mechanics were mathematically equivalent — a result whose full implications were worked out by von Neumann in the Hilbert space framework.

S16: Dirac, “The Quantum Theory of the Emission and Absorption of Radiation” (1927). This paper presents the mature formalism of quantum dynamics, introducing commutation relations and employing the Dirac delta function as a mathematical tool. It also marks the beginning of quantum electrodynamics, connecting the single-particle quantum theory to field-theoretic methods.

Mathematical Foundations

S17: von Neumann, Mathematische Grundlagen der Quantenmechanik (1932). Von Neumann’s treatise establishes quantum mechanics on rigorous Hilbert space foundations. The key contributions include the spectral theorem for unbounded self-adjoint operators, the density matrix formalism, the measurement projection postulate (which formalizes wave function collapse as a mathematical operation), and the historically important — but ultimately flawed — argument against hidden variables.

The hidden-variable argument deserves particular attention in these notes because it connects to both the Bohm entries in the wave mechanics section and the Bell entries later. Von Neumann’s proof assumes that the expectation value of a sum of observables equals the sum of the expectation values, even for observables that cannot be simultaneously measured. This linearity assumption is not physically motivated for non-commuting observables, and it is precisely this assumption that Bohm’s theory violates. Bell later made this criticism explicit and devastating.

S18: Heisenberg, The Physical Principles of the Quantum Theory (1930). Heisenberg’s monograph presents quantum mechanics from the matrix-mechanical perspective. It contains the detailed exposition of the uncertainty principle and its physical consequences, treating the uncertainty relations not merely as mathematical results but as interpretive cornerstones of the Copenhagen framework. The book is one of the primary sources for understanding what “the Copenhagen interpretation” meant to its architects, as opposed to the various simplified or caricatured versions that circulate in textbook presentations.

Group-Theoretic Foundations

S19: Weyl, The Theory of Groups and Quantum Mechanics (1931). Weyl’s text applies group-theoretic methods to quantum mechanics, developing representation theory of symmetry groups and its application to the classification of quantum states. The Weyl quantization procedure — associating classical phase-space functions with quantum operators via the Fourier transform — connects to the Moyal phase-space formulation in the next section. Weyl’s work established the structural role of symmetry in quantum theory, a perspective that would become central to modern particle physics and quantum field theory.

S20: Wigner, Group Theory and Its Application to the Quantum Mechanics of Atomic Spectra (1931). Wigner develops the role of symmetry groups in quantum mechanics with particular emphasis on rotation groups and the classification of particles by their transformation properties. This work establishes the foundations for angular momentum theory and later, for Wigner’s classification of elementary particles by representations of the Poincaré group — a thread that extends into quantum field theory and the Standard Model.

Historical Studies

S21: Pais, Niels Bohr’s Times (1991). Abraham Pais provides a detailed historical account of Bohr’s scientific career and the development of what became known as the Copenhagen interpretation. As a physicist who knew Bohr personally, Pais brings both technical understanding and firsthand perspective. The book is valuable for understanding the intellectual and institutional context in which the Copenhagen framework was established as the dominant interpretation.

S22: Cassidy, Beyond Uncertainty: The Life of Werner Heisenberg (2009). This biography examines Heisenberg’s scientific achievements within their historical context, including the development of matrix mechanics and the uncertainty principle. It complements the Pais volume by providing the Heisenberg-side perspective on the formative years of quantum mechanics.

Summary

The nine sources provide the formal backbone of quantum mechanics as it was consolidated in the late 1920s and early 1930s. The Dirac–von Neumann–Heisenberg sequence establishes the mathematical framework; the Weyl–Wigner entries add the group-theoretic perspective that would prove essential for later developments; and the biographical works supply historical context.

The section emphasizes the formalism more than on the specifically philosophical content of the Copenhagen interpretation. Bohr’s own writings — the Como lecture, the EPR reply, the collected philosophical essays — are notably absent, which means that the reader encounters the Copenhagen interpretation primarily through its mathematical codification (von Neumann, Dirac) rather than through Bohr’s own articulation of complementarity. Readers interested in these writings are referred to Pais for further consideration.

Probability, Information, and Statistics

This section addresses statistical and information-theoretic approaches to quantum mechanics, from the Born rule and ensemble interpretation through phase-space methods to modern axiomatic reconstructions. It contains seven sources spanning 1949 to 2011.

Phase-Space Methods

S23: Moyal, “Quantum Mechanics as a Statistical Theory” (1949). Moyal’s paper reformulates quantum mechanics in phase space using the Wigner function and the Moyal bracket. The Moyal bracket is a deformation of the classical Poisson bracket:

\[\{f(q,p), g(q,p)\}_{\text{Moyal}} = \frac{2}{\hbar} f(q,p) \sin \left(\frac{\hbar}{2}\left(\overleftarrow{\partial}_q \overrightarrow{\partial}_p - \overleftarrow{\partial}_p \overrightarrow{\partial}_q\right)\right) g(q,p)\]

which reduces to the Poisson bracket in the limit $\hbar \to 0$. This approach reveals quantum mechanics as a deformation of classical statistical mechanics — the noncommutativity of quantum observables is encoded in the star product rather than in the operator formalism. The phase-space formulation makes the classical limit transparent and provides a natural framework for semiclassical approximations.

The Moyal formulation has interpretive significance beyond its technical utility: it shows that one can formulate quantum mechanics without wave functions, Hilbert spaces, or operators, using instead a deformed probability distribution on phase space. This challenges the view that the Hilbert space formalism is the unique or natural mathematical setting for quantum theory.

For scientific computing, the phase-space formulation is particularly valuable. The Wigner function provides a quasiprobability distribution that can be evolved using techniques borrowed from classical statistical mechanics, making it natural for semiclassical simulations. The Wigner function can take negative values — a signature of quantum coherence — and the degree of negativity has been connected to the computational resources required for classical simulation of quantum circuits (this is the “Wigner negativity” resource theory in quantum computing). When the Wigner function is everywhere nonnegative, the corresponding quantum state can be efficiently simulated classically; it is precisely the states with negative Wigner functions that confer a quantum computational advantage. This connection between the phase-space formulation and computational complexity theory illustrates how foundational questions about the nature of the quantum state have direct computational consequences.

S29: Curtright and Zachos, “Quantum Mechanics in Phase Space” (2011; arXiv:1104.5269). This paper develops advanced mathematical techniques in phase-space quantum mechanics, focusing on the Moyal product, star products, and deformation quantization. It extends and systematizes the framework initiated by Moyal, providing the technical machinery for applications in quantum optics, quantum information, and semiclassical physics. For the reader familiar with the Moyal formulation, this paper maps out the modern state of the subject.

The Statistical Interpretation

S24: Born, “The Interpretation of Quantum Mechanics” (1953). Born’s paper articulates and defends the probabilistic interpretation of the wave function — the Born rule, which states that $\vert \psi(x) \vert^2$ gives the probability density for finding a particle at position $x$. While the Born rule was introduced much earlier (1926), this later paper represents Born’s mature philosophical defense of the statistical interpretation and its foundational role in quantum theory.

S25: Ballentine, “The Statistical Interpretation of Quantum Mechanics” (1970). Ballentine’s influential Reviews of Modern Physics article advocates treating quantum mechanics as a purely statistical theory that applies to ensembles rather than individual systems, without invoking wave function collapse. On this view, the quantum state describes the statistical properties of a preparation procedure, not the physical state of an individual system. The ensemble interpretation avoids the measurement problem — there is no collapse because there is nothing to collapse — but at the cost of leaving unanswered the question of what determines individual measurement outcomes.

Ballentine’s paper is useful for clearly distinguishing two different things that “the statistical interpretation” might mean: (1) that $\vert \psi \vert^2$ gives probabilities (the Born rule, which is accepted by essentially all interpretations), and (2) that the quantum state is a statistical description of an ensemble and nothing more (a substantive interpretive claim that not all physicists accept).

Axiomatic Reconstructions

S26: Hardy, “Quantum Theory from Five Reasonable Axioms” (2001; arXiv:quant-ph/0101012). Hardy’s paper derives quantum theory from five operationally motivated axioms:

  1. Probabilities. The state of a system is determined by the probabilities for outcomes of a set of measurements.
  2. Simplicity. Systems have the minimum information-carrying capacity consistent with distinguishability of states.
  3. Subspaces. A system that is known to pass a given test is described by a state belonging to a lower-dimensional system.
  4. Composite systems. The state of a composite system is completely determined by the statistics of measurements on the component systems.
  5. Continuity. There exist continuous reversible transformations between any two pure states.

Axioms 1–4 are satisfied by both classical and quantum probability. Axiom 5 — continuity of state transformations — is what singles out quantum theory. Dropping “continuous” from Axiom 5 recovers classical probability theory. This result is striking because it identifies a single, physically transparent condition that distinguishes quantum from classical: the existence of continuous paths in state space connecting any two pure states.

S27: Chiribella, D’Ariano, and Perinotti, “Informational Derivation of Quantum Theory” (2011; arXiv:1011.6451). This paper provides an operational, information-theoretic reconstruction of quantum theory within the framework of generalized probabilistic theories (GPTs). The key innovation is the purification postulate: every mixed state of a system can be obtained by tracing over part of a pure state of a larger system. This postulate, combined with four other axioms (causality, perfect distinguishability, ideal compression, and local distinguishability), uniquely selects quantum theory from the space of all possible probabilistic theories.

The purification postulate has a natural information-theoretic reading: it says that every source of randomness in quantum mechanics can be “purified” — traced back to entanglement with an environment. This connects to the observation that quantum randomness is fundamentally different from classical randomness: in quantum mechanics, apparent randomness at the level of a subsystem can coexist with a pure (zero-entropy) state of the whole.

S28: Banks, “Probability and Quantum Mechanics” (2011, guest post on Preposterous Universe). Tom Banks’ essay argues that quantum mechanics is an inevitable generalization of probability theory once one recognizes that observables need not commute. The essay’s central claims include: that classical logic implicitly assumes the probability sum rule $P(A \text{ or } B) = P(A) + P(B)$, which fails for quantum systems; that non-commuting observables are already implicit in classical physics via the Koopman–von Neumann formalism; and that decoherence explains the emergence of classicality by suppressing interference terms by factors of order $10^{-N}$ where $N \sim 10^{20}$ for macroscopic objects.

Summary

This section most directly engages with the “reconstruction” culture in modern quantum foundations — the program of deriving the quantum formalism from simple, physically motivated axioms. The Hardy and Chiribella papers represent the two most influential entries, and their inclusion brings the collection a contemporary character.

The phase-space methods (Moyal, Curtright–Zachos) connect to the Copenhagen section through Weyl quantization and to the wave mechanics section through the classical limit. The statistical interpretation entries (Born, Ballentine) provide the interpretive context for understanding what the reconstruction program is trying to achieve: a derivation of the formalism from principles that are interpretively transparent, rather than merely postulating the Hilbert space structure.

A connection worth noting: the Moyal formulation provides a setting in which the transition from classical to quantum is parameterized by a single deformation parameter ($\hbar$), making the classical limit a smooth process. This contrasts with the Hilbert space formalism, where the relationship between classical and quantum descriptions is more opaque. The reconstruction program can be seen as an attempt to identify the minimal physical principles that necessitate this deformation.

Bell’s Theorem

This section covers the EPR argument and Bell’s theorem, which together establish that quantum mechanics cannot be reproduced by any local hidden-variable theory. It contains four sources spanning 1935 to 1995.

The EPR Argument

S30: Einstein, Podolsky, and Rosen, “Can Quantum-Mechanical Description of Physical Reality Be Considered Complete?” (1935). The EPR paper formulates the argument that quantum mechanics is incomplete. The argument proceeds from two premises: (1) a criterion of reality — if, without disturbing a system, we can predict with certainty the value of a physical quantity, then there exists an element of physical reality corresponding to that quantity — and (2) the observation that for an entangled pair of particles, measurements on one particle allow certain prediction of measurement outcomes on the other. Since quantum mechanics does not simultaneously assign definite values to complementary observables (position and momentum, say), EPR conclude that the quantum-mechanical description is incomplete: there must be additional elements of reality not captured by the wave function.

The EPR argument does not claim that quantum mechanics is wrong, only that it is incomplete. It implicitly assumes locality — that a measurement on one particle cannot instantaneously affect the physical state of a distant particle. It is this locality assumption that Bell’s theorem will later show to be in tension with the quantum predictions.

Bell’s Analysis

S31: Bell, Speakable and Unspeakable in Quantum Mechanics (1987/2004). Bell’s collected papers on quantum foundations include his seminal 1964 paper deriving the Bell inequalities. Bell showed that any local hidden-variable theory — any theory that (a) assigns definite values to observables prior to measurement and (b) satisfies a locality condition — must obey certain statistical constraints (the Bell inequalities) that quantum mechanics predicts will be violated. Experimental tests have confirmed the quantum predictions, ruling out the class of local hidden-variable theories.

The logical structure is: EPR assumes local realism and derives incompleteness; Bell assumes local realism and derives inequalities; experiments violate the inequalities; therefore local realism fails. The conclusion is that at least one of EPR’s premises must be abandoned — either locality, or the reality criterion, or both.

Bell’s collected papers are valuable not only for the theorem itself but for his subsequent reflections on its meaning, his critiques of the Copenhagen interpretation, and his sympathetic treatment of Bohm’s pilot-wave theory (which is nonlocal but deterministic, thereby satisfying the reality criterion while violating locality).

S32: Coleman, “Quantum Mechanics In Your Face” (1994; arXiv:2011.12671, posthumously posted). Sidney Coleman’s lecture provides a characteristically clear and provocative exposition of Bell’s theorem and quantum paradoxes for a broad physics audience. Coleman’s rhetorical strategy is to convince the listener that quantum mechanics is not merely a calculational tool but a description of the world that resists classical intuition at a fundamental level. The lecture is pedagogically useful as an entry point for physicists who have learned quantum mechanics as a computational framework without engaging with its foundational implications.

S33: Peres, Quantum Theory: Concepts and Methods (1995). Asher Peres’ textbook provides a comprehensive, instrumentalist treatment of quantum foundations. The book’s approach is pragmatic, summarized in Peres’ memorable formulation: “quantum phenomena do not occur in a Hilbert space, they occur in a laboratory.” Key features include:

  • Rigorous treatment of preparations and tests as primitive notions, preceding the introduction of the Hilbert space formalism.
  • Detailed exposition of Bell’s theorem, Gleason’s theorem, and the Kochen–Specker theorem, with careful attention to what each result does and does not establish.
  • Discussion of contextuality and nonlocality as distinct concepts — a point that is often blurred in less careful treatments.
  • Treatment of quantum information, quantum thermodynamics, and quantum chaos.

Peres is placed in the Bell’s theorem section because of his extensive and influential treatment of Bell inequalities and contextuality, but the book’s scope is much broader. It functions as a foundations-and-methods reference that a reader could profitably consult alongside almost any other source in the collection.

Summary

The four sources provide the essential chain: EPR poses the problem; Bell sharpens it into a testable prediction; Coleman makes the argument accessible; Peres provides the technical apparatus in full generality. This approach to the material is compact but sufficient.

What is not represented here is the experimental side of the story — the progression from Freedman and Clauser (1972) through Aspect (1982) to the loophole-free tests of Hensen et al. (2015) and others. For a reader interested in the empirical status of Bell inequality violations, the Peres book partially fills this gap through its discussion of experimental tests. Readers interested in a more thorough account of the experimental side a referred to the primary literature.

It is worth noting the logical subtlety of Bell’s result, which is sometimes obscured in popular presentations. Bell’s theorem does not prove that quantum mechanics is nonlocal. It proves that any theory reproducing the predictions of quantum mechanics cannot be both local and realistic (in the EPR sense). Different interpretations resolve this differently: Bohm’s theory is nonlocal and realistic; the many-worlds interpretation is (arguably) local but abandons the single-outcome assumption; and various epistemic interpretations question the reality criterion itself. The Bell section of the collection provides the raw material for working through these distinctions, but the reader should be aware that the interpretive consequences of Bell’s theorem are themselves a matter of ongoing debate.

A further technical point deserves attention. Bell’s original inequality applies to a specific experimental setup (measurements of spin-1/2 particles along two directions each). The generalization to the CHSH inequality (Clauser, Horne, Shimony, and Holt, 1969) provides a more experimentally accessible bound, and Tsirelson’s bound ($2\sqrt{2}$) gives the maximum quantum violation of the CHSH inequality. The gap between the classical bound (2), the quantum bound ($2\sqrt{2} \approx 2.83$), and the algebraic maximum (4) has been the subject of extensive investigation, connecting Bell’s theorem to information-theoretic principles (information causality, macroscopic locality) and to the structure of the space of correlations in generalized probabilistic theories.

Everett’s Many Worlds

This section covers the relative-state (many-worlds) interpretation, which resolves the measurement problem by eliminating wave function collapse from the theory. It contains four sources, all closely connected, spanning 1957 to 2012.

The Original Formulation

S34: Everett, “‘Relative State’ Formulation of Quantum Mechanics” (1957). Hugh Everett’s Reviews of Modern Physics paper introduces the relative-state formulation. The central proposal is that wave function collapse never occurs: the universal wave function always evolves unitarily according to the Schrödinger equation. What appears to be “collapse” is reinterpreted as the establishment of correlations (entanglement) between the observer and the observed system. After a measurement interaction, the observer does not have a single definite outcome; rather, different “branches” of the universal wave function contain different observer records correlated with different measurement results.

The key technical move is to replace the collapse postulate with the claim that all branches of the superposition are equally real. The apparent definiteness of measurement outcomes is explained by the fact that each branch contains an observer who records a definite result — the branching structure of the wave function mimics the phenomenology of collapse without requiring it.

S35: Wheeler, “Assessment of Everett’s ‘Relative State’ Formulation of Quantum Mechanics” (1957). John Wheeler’s companion paper provides a critical evaluation of Everett’s approach. As Everett’s doctoral advisor, Wheeler was in a unique position to assess the work. His paper discusses the philosophical and physical implications while expressing both enthusiasm for the formalism and caution about its ontological commitments. Wheeler’s assessment is historically significant as the earliest serious engagement with Everett’s ideas by a physicist of the first rank.

Development and Analysis

S36: DeWitt, “The Many-Worlds Interpretation of Quantum Mechanics” (1973). Bryce DeWitt’s work expands and popularizes the many-worlds interpretation, coining the vivid “many-worlds” terminology that replaced Everett’s more austere “relative state” language. DeWitt addresses early criticisms and develops the concept of branching into distinct “worlds.” The shift in terminology — from “relative state” to “many worlds” — is not merely cosmetic: it reflects a different emphasis, from the formal structure of the theory (relative states within a single Hilbert space) to an ontological picture (a constantly branching multiverse).

S37: Barrett and Byrne, eds., The Everett Interpretation of Quantum Mechanics: Collected Works 1955–1980 with Commentary (2012). This Princeton volume provides modern, comprehensive analysis of the Everett interpretation. It collects Everett’s original writings (including unpublished material), Wheeler’s correspondence and assessments, and critical commentary addressing the major open problems: the preferred basis problem (what determines the branching structure?), the probability problem (how does the Born rule emerge from deterministic unitary evolution?), and the ontological status of the branches.

Summary

The four sources trace a minimal but canonical path from Everett’s original formulation through Wheeler’s early assessment and DeWitt’s popularization to modern philosophical and technical analysis. This is a minimal introduction. Note that a significant body of technical work has accumulated since the 1990s on the preferred basis problem and the probability problem. Zurek’s work on decoherence and “environment-assisted invariance” (envariance), the decision-theoretic approaches of Deutsch and Wallace, and Saunders’ philosophical analysis substantially expand on the original view from the 1950’s. The Barrett–Byrne volume provides pointers to some of this work, but dedicated sources are beyond the scope considered here.

The probability problem deserves particular attention because it connects to both the foundations of probability theory and to practical questions in quantum computation. In a deterministic, unitarily-evolving universe, what does it mean to say that an outcome has probability $\vert\alpha\vert^2$? The Deutsch–Wallace decision-theoretic approach argues that a rational agent who believes in the Everett interpretation is compelled to set their credences equal to the Born-rule probabilities, on pain of Dutch-book arguments. This approach has been influential but remains controversial; the main objection is that it smuggles in assumptions about rationality that do the work usually done by the Born rule itself.

For computational purposes, the Everettian perspective suggests viewing quantum computation as a process that operates simultaneously on all branches of the wave function. This is heuristically useful — it motivates the idea that quantum computers achieve speedups by exploiting parallel computation across branches — but it is also potentially misleading, since the interference structure that enables quantum speedups is more subtle than simple parallelism. The tensor network perspective on many-body quantum states provides a more precise computational framework for understanding the branching structure of quantum evolution.

Gaps

The collection here is an idiosyncratic sampling. Several major interpretive programs of the 20th and 21st centuries are not represented. This section documents the principal gaps and provides entry points for readers who wish to extend the collection.

Spontaneous Collapse Theories

The GRW (Ghirardi–Rimini–Weber, 1986) theory and its continuous generalization CSL (continuous spontaneous localization; Pearle, 1989) modify the Schrödinger equation by adding stochastic, nonlinear terms that cause superpositions of macroscopically distinct states to collapse spontaneously. These theories make the same predictions as standard quantum mechanics for microscopic systems but differ for macroscopic superpositions, making them in principle empirically testable. The absence of collapse theories from the collection is a significant gap: they represent the only interpretive program that modifies the dynamics of quantum mechanics (as opposed to reinterpreting the existing formalism).

The GRW mechanism introduces two new physical constants: a collapse rate $\lambda \approx 10^{-16}$ s$^{-1}$ per particle and a localization width $r_C \approx 10^{-7}$ m. For a single particle, collapses are extremely rare; for a macroscopic object containing $N \sim 10^{23}$ particles, the effective collapse rate is $N\lambda \sim 10^7$ s$^{-1}$, ensuring that macroscopic superpositions are destroyed almost instantaneously. The CSL extension makes the collapse process continuous rather than discrete, resulting in a stochastic nonlinear modification of the Schrödinger equation that has the structure of a stochastic differential equation driven by a Wiener process. This makes collapse theories amenable to the standard tools of stochastic calculus and numerical simulation.

Current experimental efforts to test collapse models focus on mechanical oscillators, matter-wave interferometry, and spontaneous radiation from charged particles. The predicted departures from standard quantum mechanics are extremely small, but improving experimental sensitivity may eventually provide empirical discrimination between collapse and no-collapse theories.

Entry points: Ghirardi, Rimini, and Weber, “Unified dynamics for microscopic and macroscopic systems,” Physical Review D 34, 470 (1986); Bassi and Ghirardi, “Dynamical reduction models,” Physics Reports 379, 257 (2003); Bassi et al., “Models of wave-function collapse, underlying theories, and experimental tests,” Reviews of Modern Physics 85, 471 (2013).

QBism

QBism (Quantum Bayesianism) interprets quantum states as expressions of an agent’s personal beliefs about future measurement outcomes, updated via a quantum generalization of Bayesian conditioning. The Born rule becomes a normative constraint on rational belief rather than a law of nature. QBism dissolves the measurement problem by denying that quantum states are objective features of the world.

Entry points: Fuchs, Mermin, and Schack, “An introduction to QBism with an application to the locality of quantum mechanics,” American Journal of Physics 82, 749 (2014); Caves, Fuchs, and Schack, “Quantum probabilities as Bayesian probabilities,” Physical Review A 65, 022305 (2002).

Consistent (Decoherent) Histories

The consistent histories framework (Griffiths, 1984; Omnès, 1988; Gell-Mann and Hartle, 1990) provides rules for assigning probabilities to sequences of quantum events (“histories”) without invoking wave function collapse. A set of histories is “consistent” (or “decoherent”) if the interference terms between distinct histories vanish, permitting a classical probability calculus. The framework is used extensively in quantum cosmology, where the absence of an external observer makes collapse-based formulations problematic.

Entry points: Griffiths, Consistent Quantum Theory (Cambridge, 2002); Gell-Mann and Hartle, “Classical equations for quantum systems,” Physical Review D 47, 3345 (1993).

Decoherence

Decoherence is not itself an interpretation but a physical process — the suppression of interference through entanglement with the environment — that plays a role in several interpretive frameworks (Everettian, consistent histories, and others). The collection’s sources reference decoherence in passing (Banks, Landsman) but do not include a dedicated treatment.

Entry points: Zurek, “Decoherence, einselection, and the quantum origins of the classical,” Reviews of Modern Physics 75, 715 (2003); Schlosshauer, Decoherence and the Quantum-to-Classical Transition (Springer, 2007).

Suggested Reading Paths

The 37 sources in the collection can be approached in several ways, depending on the reader’s background and interests. The following paths are suggestions, not prescriptions.

Path: Historical. For a reader interested in the development of quantum mechanics as an intellectual enterprise: S1 (de Broglie) $\to$ S2 (Schrödinger) $\to$ S14–S16 (Dirac) $\to$ S18 (Heisenberg) $\to$ S17 (von Neumann) $\to$ S30 (EPR) $\to$ S3–S4 (Bohm) $\to$ S34 (Everett) $\to$ S31 (Bell). Supplement with the biographical works (S21 Pais, S22 Cassidy) and the Solvay reconstruction (S8 Bacciagaluppi–Valentini).

Path: Formal/Mathematical. For a reader interested in the mathematical structure: S17 (von Neumann) $\to$ S19 (Weyl) $\to$ S20 (Wigner) $\to$ S23 (Moyal) $\to$ S29 (Curtright–Zachos) $\to$ S33 (Peres) $\to$ S31 (Bell) $\to$ S27 (Chiribella).

Path: Interpretive. For a reader interested in the interpretation debate: S30 (EPR) $\to$ S31 (Bell) $\to$ S32 (Coleman) $\to$ S3–S4 (Bohm) $\to$ S34 (Everett) $\to$ S25 (Ballentine) $\to$ S11 (PBR) $\to$ S26 (Hardy) $\to$ S33 (Peres).

Path: Reconstruction. For a reader interested in the information-theoretic reconstruction program: S24 (Born) $\to$ S25 (Ballentine) $\to$ S23 (Moyal) $\to$ S26 (Hardy) $\to$ S27 (Chiribella) $\to$ S11 (PBR) $\to$ S28 (Banks).

Path: Computational. For a reader approaching foundations from the perspective of scientific computing: S17 (von Neumann) — Hilbert space structure and measurement postulates $\to$ S23 (Moyal) — phase-space methods and Wigner function $\to$ S29 (Curtright–Zachos) — deformation quantization machinery $\to$ S3–S4 (Bohm) — trajectory methods $\to$ S31 (Bell) — nonlocality constraints $\to$ S27 (Chiribella) — axiomatic structure of quantum theory within GPTs $\to$ S13 (Landsman) — measurement problem and its computational implications. Supplement with the decoherence literature (Zurek, Schlosshauer) for the connection to classical simulation.

Bibliography

Wave Mechanics

IDLinkNotes
S1deBroglie “On the Theory of Quanta” 1925Introduces the concept of matter waves and posits wave–particle duality for quantum entities.
S2Schrödinger “Collected Papers” 1928Compiles Schrödinger’s foundational papers outlining the wave equation formulation of quantum mechanics.
S3Bohm “A Suggested Interpretation I” 1951Proposes a hidden‐variable interpretation of quantum mechanics, introducing the pilot‐wave concept.
S4Bohm “A Suggested Interpretation II” 1951Expands on the hidden‐variable framework and the detailed dynamics of the pilot wave in quantum systems.
S5Schrödinger “Jumps I” 1952Analyzes the notion of quantum jumps and their reconciliation with a continuous wave picture.
S6Schrödinger “Jumps II” 1952Continues the examination of quantum discontinuities, exploring how wave mechanics can accommodate “jumps.”
S7Wallstrom “Schrödinger’s & Madelung’s Equations” 1994Investigates the equivalence and differences between Schrödinger’s equation and the Madelung hydrodynamic formulation.
S8Bacciagaluppi and Valentini “Quantum Theory at the Crossroads” 2006Provides a historical and philosophical survey of competing interpretations of quantum mechanics during the 1927 Solvay Conference.
S9N. A. Wheeler “Schrödinger’s Argument” 2006Revisits Schrödinger’s original approach to wave mechanics, illustrating the optical mechanical analogy.
S10Joas and Lehner “Optical Mechanical Analogy” 2009Explores the analogy between geometric optics and classical mechanics as a precursor to wave‐mechanical insights.
S11PBR “Reality of the Quantum State” 2012Demonstrates via a no‐go theorem that the quantum state must correspond to an element of reality rather than mere information.
S12Lopes Coelho and Stachel “Schrödinger, Hertz, and Van Vleck” 2013Examines historical influences on Schrödinger’s formulation, particularly drawing from Hertz’s and Van Vleck’s work.
S13Landsman “Measurement” 2017Discusses the measurement problem in quantum theory, tracing its development and formal challenges.

Copenhagen Interpretation

IDLinkNotes
S14Dirac “Fundamental Equations” 1925Derives core equations of quantum theory, laying groundwork for matrix mechanics and quantization.
S15Dirac “Theory of Quantum Mechanics” 1926Introduces the transformation theory, formalizing wavefunction collapse and operator methods in quantum mechanics.
S16Dirac “Quantum Dynamics” 1927Presents the formalism of quantum dynamics, including commutation relations and the Dirac delta function.
S17von Neumann “Mathematical Foundations” 1929Establishes a rigorous Hilbert‐space formulation underpinning the statistical interpretation of quantum theory.
S18Heisenberg “Physical Principles of Quantum Theory” 1930Details the matrix‐mechanics framework, introduces the uncertainty principle, and connects observables to operators.
S19Weyl “The Theory of Groups and Quantum Mechanics” 1931Applies group‐theoretic methods to classify quantum states and symmetries, influencing representation theory.
S20Wigner “Group Theory” 1931Develops the role of symmetry groups in quantum mechanics, especially for particle classification.
S21Pais “Niels Bohr’s Times” 1991Provides a detailed historical account of Bohr’s development of the Copenhagen interpretation and related debates.
S22Cassidy “Beyond Uncertainty” 2009Provides a detailed biography of Werner Heisenberg, examining his scientific achievements, personal life, and the historical context of quantum mechanics.

Probability, Information, Statistics

IDLinkNotes
S23Moyal “Quantum Mechanics as a Statistical Theory” 1949Shows how quantum mechanics can be reformulated in phase space using a statistical framework (Moyal bracket).
S24Born “Interpretation of Quantum Mechanics” 1953Articulates the probabilistic interpretation of the wavefunction (Born rule) and its foundational role in quantum theory.
S25Ballentine “Statistical Interpretations” 1970Advocates treating quantum mechanics purely as a statistical theory without invoking wavefunction collapse as physical.
S26Hardy “5 Reasonable Axioms” 2001Derives the formal structure of quantum theory from five intuitive axioms, emphasizing informational and probabilistic aspects.
S27Chiribella “Informational Derivation” 2011Provides an operational, information‐theoretic reconstruction of quantum theory based on principles of information processing.
S28Banks “Probability and Quantum Mechanics” 2011Discusses various interpretational issues in quantum mechanics, with emphasis on probability and the decoherence view.
S29Curtright and Zachos “Moyal Formalism” 2011Develops mathematical techniques in phase‐space quantum mechanics, focusing on the Moyal product and deformation quantization.

Bell’s Theorem

IDLinkNotes
S30Einstein Podolsky Rosen “Can QM be Complete?” 1935Formulates the EPR paradox, arguing that quantum mechanics is incomplete due to apparent “spooky action at a distance.”
S31Bell “Speakable and Unspeakable” 1989Collects Bell’s key papers on nonlocality, culminating in the Bell inequalities that test hidden‐variable theories.
S32Coleman “Quantum Mechanics In Your Face” 1994Offers a clear, informal critique of quantum mechanical paradoxes, including discussion of Bell’s theorem for broader audiences.
S33Peres “Concepts and Methods”Provides a comprehensive introduction to quantum information concepts and methods, including detailed treatment of Bell’s theorem.

Everett’s “Many Worlds”

IDLinkNotes
S34Everett “Relative State” 1957Introduces the relative‐state (many‐worlds) interpretation, proposing that all possible outcomes occur in branching universes.
S35Wheeler “Assessment of Everett’s Relative State” 1957Critically evaluates Everett’s approach, discussing its philosophical and physical implications for quantum measurement.
S36DeWitt “Many Worlds” 1973Expands and popularizes the many‐worlds interpretation through detailed exposition and responses to early criticisms.
S37Barrett and Byrne “The Everett Interpretation” 2012Provides a modern, comprehensive analysis of Everett’s interpretation, examining both formal developments and philosophical debates.

This post is licensed under CC BY 4.0 by the author.