Note: This blog post is inspired by my own term paper for MIT 8.06x (Applications of Quantum Mechanics). You can find the original PDF here: Case Study of the Momentum Operator.
Why this post exists
Most quantum mechanics textbooks tell you that observables correspond to “Hermitian operators.” This is, charitably, a half-truth — and the half that’s missing is exactly the half that gets you into trouble the moment you try to define a momentum operator on anything more interesting than .
The issue is that “Hermitian” in the physics sense usually means symmetric, but what observables really need to be is self-adjoint, and these are not the same thing in infinite dimensions. The gap between them is invisible in finite-dimensional linear algebra (where every linear operator is bounded, so every densely defined symmetric operator is automatically self-adjoint). In quantum mechanics, where Hilbert spaces are typically and the operators we care about are unbounded, the distinction is unavoidable.
This post is a case study. We’ll work through the definitions carefully — operator, adjoint, symmetric, self-adjoint, essentially self-adjoint — and then apply them to the momentum operator in two settings:
- A compact interval with hard-wall (Dirichlet) boundary conditions.
- A circle (the same interval with periodic boundary conditions).
The punchline: on the circle, the momentum operator is essentially self-adjoint and there’s a unique sensible “promotion” of it to a true observable. On the interval with Dirichlet boundary conditions, it is not essentially self-adjoint, and the question of which self-adjoint operator deserves to be called “the momentum on ” doesn’t have a unique answer — there’s a one-parameter family of options, and you have to choose. That’s the kind of subtlety that gets papered over when you treat as if it were a finite matrix.
I’ll assume a working knowledge of Hilbert spaces and basic functional analysis (bounded operators, dense subspaces, ). Set throughout.
1. Operators, bounded and unbounded
An operator between two normed spaces and is just a linear map . Nothing special so far. The interesting part is the topology.
In finite dimensions, every linear operator is automatically continuous, because all norms on a finite-dimensional vector space are equivalent. This is false in infinite dimensions, and the failure is not a mathematical curiosity — it’s the reason quantum mechanics needs a more careful framework than ordinary linear algebra.
Bounded operators
Let be a normed space and a Banach space. A linear operator is bounded if
equivalently if there exists a constant such that for all .
For linear operators, “bounded” and “continuous” are synonymous (this is a standard lemma). We write for the space of bounded operators .
Unbounded operators
A linear operator that fails the boundedness condition is called unbounded. The two most important operators in quantum mechanics — position and momentum — are both unbounded on :
- Position: .
- Momentum: .
Why are they unbounded? Position because you can find functions concentrated arbitrarily far from the origin where blows up. Momentum because differentiation amplifies high-frequency components without bound.
A more elementary way to see that the derivative operator is unbounded: define
Take , so but . The ratio is unbounded.
Important consequence. An unbounded operator cannot be defined on the whole Hilbert space — at least not as a closed linear operator (this is the Hellinger–Toeplitz theorem). It always lives on a proper subspace called its domain. Specifying that domain is part of specifying the operator. Two operators with the same formal expression but different domains are different operators. This is the single most important thing to internalize.
2. Adjoints, symmetric, and self-adjoint
Densely defined operators
Let be a Hilbert space. A linear operator is densely defined if its domain is dense in , i.e. every vector in can be approximated arbitrarily well by vectors in .
Density is what allows us to define the adjoint at all. Without it, the adjoint is not even well-defined as a single-valued operator.
The adjoint
Let be densely defined. The adjoint is the operator with domain
acting by .
is well-defined. If both and satisfy for all , then for all in a dense set, which forces . (This is exactly where density is used.)
A small but useful proposition: . To see it, for all .
Extensions and a key inclusion lemma
We say is an extension of , written , if and for all .
Lemma. If are both densely defined, then .
Proof. Let , so there exists with for all . Since and on , this gives for all . Hence and .
Notice the direction-reversal: a bigger operator has a smaller adjoint. This is the crucial structural fact.
Symmetric vs. self-adjoint
A densely defined operator is symmetric if
It is self-adjoint if — meaning both the actions agree and the domains agree:
- ,
- for all .
A linguistic warning. In physics, the word “Hermitian” is used inconsistently: sometimes as a synonym for symmetric, sometimes as a synonym for self-adjoint. Statements like “observables correspond to Hermitian operators” gloss over a real distinction that becomes invisible only in finite dimensions. I’ll avoid the word entirely.
If is symmetric, then . (For , set ; symmetry gives for all , so with .) The whole game is to ask whether the inclusion is equality.
A maximality property
Self-adjoint operators have no proper symmetric extensions. Concretely: if is self-adjoint, is symmetric, and , then .
Proof. From , the lemma gives . Symmetric means , so . Combined with , we get .
This is why self-adjointness is the right condition for observables. Spectral theorem guarantees self-adjoint operators have real spectra and a sensible functional calculus; symmetric operators in general do not. And the maximality property tells us that a self-adjoint operator is “as big as it can be” — there’s no room to extend it further while keeping the symmetric property.
3. Closability, closure, essentially self-adjoint
There’s one more layer we need. Even if our operator isn’t self-adjoint, it may have a unique self-adjoint extension, hidden inside . This is captured by the notion of essentially self-adjointness.
Closures
A densely defined operator is closable if its adjoint is itself densely defined. Equivalently (and more usefully), has a smallest closed extension, which we call the closure . It satisfies .
Symmetric operators are always closable. If is symmetric, then implies . Since is dense, so is , hence is closable. So we always have access to for symmetric operators.
For symmetric , the closure sits between and :
Essentially self-adjoint operators
A symmetric operator is essentially self-adjoint (e.s.a.) if its closure is self-adjoint.
This is weaker than self-adjointness. A self-adjoint operator is automatically e.s.a.: if , then taking adjoints gives , so , so is self-adjoint.
Theorem. If is essentially self-adjoint, then is the unique self-adjoint extension of .
Proof. is a self-adjoint extension by hypothesis. For uniqueness, suppose is any other self-adjoint extension, . Self-adjoint operators are closed (since and adjoints are always closed), so (the closure is the smallest closed extension). But is itself self-adjoint, and a self-adjoint operator has no proper symmetric extensions — so .
The strategic upshot. When we want to elevate a symmetric operator to a genuine observable, we do not need it to already be self-adjoint. It suffices that it be essentially self-adjoint. Then the closure is the canonical observable, picked out uniquely.
When essential self-adjointness fails, life gets harder: the symmetric operator may admit many self-adjoint extensions (or even none), and choosing one becomes a piece of physical input. The technology to count and parametrize these extensions is deficiency index theory, due to von Neumann — beyond our scope here, but coming up explicitly when we look at the interval case.
4. The momentum operator: setup
Now to the main event. Define the momentum operator on the -th coordinate as
where the domain has yet to be specified. We work on in one dimension, and we’ll consider two natural-looking domains:
- Interval (Dirichlet): .
- Circle (periodic): .
The Dirichlet condition is strictly stronger than the periodic one ( implies , but not conversely), so as operators we have — the circle operator is an extension of the interval operator.
We’ll need a couple of function-space facts. The chain of inclusions
relates classically differentiable functions () to absolutely continuous functions () and the Sobolev space . Recall:
- if there exists a Lebesgue-integrable such that , in which case almost everywhere.
- .
The point of is that it’s the natural domain for ” has an derivative,” allowing us to apply and stay in — without requiring classical differentiability everywhere.
5. Momentum on the interval
Take with domain .
Step 1. is symmetric
For , integrate by parts:
The boundary term vanishes because . The remaining integral is
So is symmetric. ✓
Step 2. The adjoint has a bigger domain
Now we ask: for which does there exist with
The strategy: pick any with a.e. (such an exists because every function on a compact interval has an absolutely continuous antiderivative). Substituting and integrating the right-hand side by parts,
and the boundary term vanishes because vanishes at the endpoints. So becomes
Equivalently, in .
Now we identify this orthogonal complement. The set consists exactly of those continuous functions on satisfying — that is, , where is the constant function. (One direction: if with , then . Conversely, given such a , the antiderivative vanishes at both endpoints and lies in .)
So the closure (in ) of is , and its orthogonal complement is therefore , the space of constant functions. Hence
Since , this gives with no boundary conditions whatsoever. The requirement then forces , i.e. . Conversely, every function arises this way, and so
The adjoint is the same differential expression, but acting on a much larger domain — functions with no boundary conditions, versus functions vanishing at both endpoints.
In particular, , so is not self-adjoint.
Step 3. The closure is also too small
Maybe the closure rescues us? Let’s compute directly.
means: there exists such that for all . Since and both act as , , and the question reduces to whether the appropriate boundary terms vanish.
For (which we already know is necessary, since ) and , integration by parts gives
Requiring this to vanish for all — and functions can take arbitrary independent values at and — forces and .
So
which is strictly smaller than . Hence , and is not essentially self-adjoint.
What’s really going on
This is not a defect of our choice of “starting domain” — it reflects a real physical ambiguity. There is in fact a one-parameter family of self-adjoint extensions of , parametrized by , with domains
Different choices of correspond to physically different theories (think: a particle on an interval threaded by a magnetic flux, or a “twisted” boundary condition). The deficiency-index analysis confirms that this is the full story: the deficiency indices of are , giving exactly a -worth of self-adjoint extensions.
The lesson: when an operator is symmetric but not essentially self-adjoint, the physics of which observable to use is not determined by the formal expression alone. You have to add a choice of boundary condition, and that choice is part of defining the system.
6. Momentum on the circle
Now repeat the analysis with the milder boundary condition (no requirement that the common value be zero).
Step 1. Symmetric
For with and , the boundary term in integration by parts is
using both boundary conditions. So is symmetric. ✓
Step 2. The adjoint
We can shortcut the computation by using the inclusion . By the inclusion lemma, . So and the action is still .
To pin down the boundary conditions: requires the boundary term to vanish for all with :
Since is arbitrary, this forces . So
The adjoint inherits the same periodic boundary condition, just on a bigger function space.
Step 3. Not self-adjoint, but…
Comparing,
So is not self-adjoint either — but only because of the regularity gap , not because of any boundary-condition mismatch.
Step 4. Essentially self-adjoint!
Let’s compute the closure. For and , the boundary term is
(using ). For this to vanish for all such , we need — but no further constraint, because is the only free quantity. So
Therefore , and is essentially self-adjoint.
Conclusion
The momentum operator on the circle has a unique self-adjoint realization: take the closure of the naive operator on -with-periodic-BC, and you land on the operator with domain .
This is the operator whose eigenfunctions are with eigenvalues , which is exactly the spectrum of momentum on that you’d compute formally in any QM textbook. The point of the whole apparatus is to justify that calculation by showing that the operator we’re diagonalizing actually exists as a genuine observable.
7. The takeaway
| Setting | Symmetric? | Self-adjoint? | Essentially self-adjoint? | Self-adjoint extensions |
|---|---|---|---|---|
| Interval , | yes | no | no | one-parameter family |
| Circle, | yes | no (regularity only) | yes | unique: the closure |
The interval case is a subtle example of a phenomenon that recurs throughout quantum mechanics: a perfectly innocent-looking differential operator may fail to define a unique observable, because the boundary conditions encode physical information that cannot be recovered from the formal expression alone. The circle case is the happier scenario where everything works out, and it works out because the topology of the configuration space already does the boundary-condition-fixing for us.
The general technology to handle the harder cases — counting the deficiency indices to find a -family of self-adjoint extensions when , or no extensions at all when — is due to von Neumann. It’s the right machinery for asking systematically “when is this symmetric operator a genuine observable, and if not, what data is missing?”
We also haven’t said anything about why momentum is in the first place. That’s the content of the Stone–von Neumann theorem, which from the canonical commutation relation (and an irreducibility hypothesis) recovers the Schrödinger representation up to unitary equivalence. The classical-mechanics analogue is the Poisson bracket ; quantization replaces Poisson brackets by commutators, and Stone–von Neumann tells you that this prescription has, up to mathematical niceness conditions, essentially one realization. That’s a separate story for another post.
References
- B. C. Hall, Quantum Theory for Mathematicians, Springer (2013) — Chapters 7–10.
- M. Reed and B. Simon, Methods of Modern Mathematical Physics, Volume I: Functional Analysis, Academic Press (1980) — especially Chapters VI–VIII.
- F. Schuller, Lectures on Quantum Theory, University of Erlangen-Nürnberg, YouTube playlist (Lectures 6–9 cover this material).
- P. Szekeres, A Course in Modern Mathematical Physics: Groups, Hilbert Space and Differential Geometry, Cambridge University Press (2004) — Chapters 13–14 for background reading.