3133 words
16 minutes
When 'Hermitian' Isn't Enough: A Case Study of the Momentum Operator

Note: This blog post is inspired by my own term paper for MIT 8.06x (Applications of Quantum Mechanics). You can find the original PDF here: Case Study of the Momentum Operator.

Why this post exists#

Most quantum mechanics textbooks tell you that observables correspond to “Hermitian operators.” This is, charitably, a half-truth — and the half that’s missing is exactly the half that gets you into trouble the moment you try to define a momentum operator on anything more interesting than Rn\mathbb{R}^n.

The issue is that “Hermitian” in the physics sense usually means symmetric, but what observables really need to be is self-adjoint, and these are not the same thing in infinite dimensions. The gap between them is invisible in finite-dimensional linear algebra (where every linear operator is bounded, so every densely defined symmetric operator is automatically self-adjoint). In quantum mechanics, where Hilbert spaces are typically L2(Rn)L^2(\mathbb{R}^n) and the operators we care about are unbounded, the distinction is unavoidable.

This post is a case study. We’ll work through the definitions carefully — operator, adjoint, symmetric, self-adjoint, essentially self-adjoint — and then apply them to the momentum operator P^=ix\hat{P} = -i\hbar\,\partial_x in two settings:

  1. A compact interval [0,2π][0, 2\pi] with hard-wall (Dirichlet) boundary conditions.
  2. A circle (the same interval with periodic boundary conditions).

The punchline: on the circle, the momentum operator is essentially self-adjoint and there’s a unique sensible “promotion” of it to a true observable. On the interval with Dirichlet boundary conditions, it is not essentially self-adjoint, and the question of which self-adjoint operator deserves to be called “the momentum on [0,2π][0,2\pi]” doesn’t have a unique answer — there’s a one-parameter family of options, and you have to choose. That’s the kind of subtlety that gets papered over when you treat ix-i\hbar\,\partial_x as if it were a finite matrix.

I’ll assume a working knowledge of Hilbert spaces and basic functional analysis (bounded operators, dense subspaces, L2L^2). Set =1\hbar = 1 throughout.


1. Operators, bounded and unbounded#

An operator between two normed spaces AA and BB is just a linear map T:ABT: A \to B. Nothing special so far. The interesting part is the topology.

In finite dimensions, every linear operator is automatically continuous, because all norms on a finite-dimensional vector space are equivalent. This is false in infinite dimensions, and the failure is not a mathematical curiosity — it’s the reason quantum mechanics needs a more careful framework than ordinary linear algebra.

Bounded operators#

Let (V,V)(V, \|\cdot\|_V) be a normed space and (W,W)(W, \|\cdot\|_W) a Banach space. A linear operator A:VWA: V \to W is bounded if

supfV{0}AfWfV<,\sup_{f \in V \setminus \{0\}} \frac{\|Af\|_W}{\|f\|_V} < \infty,

equivalently if there exists a constant C0C \geq 0 such that AxWCxV\|Ax\|_W \leq C\|x\|_V for all xVx \in V.

For linear operators, “bounded” and “continuous” are synonymous (this is a standard lemma). We write B(H)\mathcal{B}(\mathcal{H}) for the space of bounded operators HH\mathcal{H} \to \mathcal{H}.

Unbounded operators#

A linear operator that fails the boundedness condition is called unbounded. The two most important operators in quantum mechanics — position and momentum — are both unbounded on L2(Rn)L^2(\mathbb{R}^n):

  • Position: x^ψ(x)=xψ(x)\hat{x}\psi(x) = x\psi(x).
  • Momentum: p^ψ(x)=ixψ(x)\hat{p}\psi(x) = -i\,\partial_x\psi(x).

Why are they unbounded? Position because you can find L2L^2 functions concentrated arbitrarily far from the origin where xψ/ψ\|x\psi\|/\|\psi\| blows up. Momentum because differentiation amplifies high-frequency components without bound.

A more elementary way to see that the derivative operator is unbounded: define

D:C1[0,1]C0[0,1],ff.D: C^1[0,1] \to C^0[0,1], \qquad f \mapsto f'.

Take fn(x)=sin(nπx)f_n(x) = \sin(n\pi x), so fn=1\|f_n\|_\infty = 1 but fn=nπ\|f_n'\|_\infty = n\pi \to \infty. The ratio is unbounded.

Important consequence. An unbounded operator cannot be defined on the whole Hilbert space — at least not as a closed linear operator (this is the Hellinger–Toeplitz theorem). It always lives on a proper subspace called its domain. Specifying that domain is part of specifying the operator. Two operators with the same formal expression but different domains are different operators. This is the single most important thing to internalize.


2. Adjoints, symmetric, and self-adjoint#

Densely defined operators#

Let H\mathcal{H} be a Hilbert space. A linear operator T:DTHT: \mathcal{D}_T \to \mathcal{H} is densely defined if its domain DT\mathcal{D}_T is dense in H\mathcal{H}, i.e. every vector in H\mathcal{H} can be approximated arbitrarily well by vectors in DT\mathcal{D}_T.

Density is what allows us to define the adjoint at all. Without it, the adjoint is not even well-defined as a single-valued operator.

The adjoint#

Let T:DTHT: \mathcal{D}_T \to \mathcal{H} be densely defined. The adjoint TT^* is the operator with domain

DT:={ψHηH such that ψTφ=ηφ for all φDT},\mathcal{D}_{T^*} := \big\{ \psi \in \mathcal{H} \,\big|\, \exists\,\eta \in \mathcal{H}\ \text{such that}\ \langle\psi | T\varphi\rangle = \langle\eta|\varphi\rangle \text{ for all } \varphi \in \mathcal{D}_T \big\},

acting by Tψ:=ηT^*\psi := \eta.

TT^* is well-defined. If both η\eta and η~\tilde\eta satisfy ψTφ=ηφ=η~φ\langle\psi|T\varphi\rangle = \langle\eta|\varphi\rangle = \langle\tilde\eta|\varphi\rangle for all φDT\varphi \in \mathcal{D}_T, then ηη~φ=0\langle\eta - \tilde\eta | \varphi\rangle = 0 for all φ\varphi in a dense set, which forces η=η~\eta = \tilde\eta. (This is exactly where density is used.)

A small but useful proposition: ker(T)=ran(T)\ker(T^*) = \operatorname{ran}(T)^\perp. To see it, ψker(T)    Tψ=0    ψTφ=0φ=0\psi \in \ker(T^*) \iff T^*\psi = 0 \iff \langle\psi|T\varphi\rangle = \langle 0|\varphi\rangle = 0 for all φDT    ψran(T)\varphi \in \mathcal{D}_T \iff \psi \perp \operatorname{ran}(T).

Extensions and a key inclusion lemma#

We say T~\tilde T is an extension of TT, written TT~T \subseteq \tilde T, if DTDT~\mathcal{D}_T \subseteq \mathcal{D}_{\tilde T} and T~φ=Tφ\tilde T\varphi = T\varphi for all φDT\varphi \in \mathcal{D}_T.

Lemma. If TT~T \subseteq \tilde T are both densely defined, then T~T\tilde T^* \subseteq T^*.

Proof. Let ψDT~\psi \in \mathcal{D}_{\tilde T^*}, so there exists ηH\eta \in \mathcal{H} with ψT~β=ηβ\langle\psi|\tilde T\beta\rangle = \langle\eta|\beta\rangle for all βDT~\beta \in \mathcal{D}_{\tilde T}. Since DTDT~\mathcal{D}_T \subseteq \mathcal{D}_{\tilde T} and T~=T\tilde T = T on DT\mathcal{D}_T, this gives ψTα=ηα\langle\psi|T\alpha\rangle = \langle\eta|\alpha\rangle for all αDT\alpha \in \mathcal{D}_T. Hence ψDT\psi \in \mathcal{D}_{T^*} and Tψ=η=T~ψT^*\psi = \eta = \tilde T^*\psi. \square

Notice the direction-reversal: a bigger operator has a smaller adjoint. This is the crucial structural fact.

Symmetric vs. self-adjoint#

A densely defined operator TT is symmetric if

αTβ=Tαβfor all α,βDT.\langle \alpha | T\beta\rangle = \langle T\alpha | \beta \rangle \qquad \text{for all } \alpha, \beta \in \mathcal{D}_T.

It is self-adjoint if T=TT = T^* — meaning both the actions agree and the domains agree:

  1. DT=DT\mathcal{D}_T = \mathcal{D}_{T^*},
  2. Tφ=TφT\varphi = T^*\varphi for all φDT\varphi \in \mathcal{D}_T.

A linguistic warning. In physics, the word “Hermitian” is used inconsistently: sometimes as a synonym for symmetric, sometimes as a synonym for self-adjoint. Statements like “observables correspond to Hermitian operators” gloss over a real distinction that becomes invisible only in finite dimensions. I’ll avoid the word entirely.

If TT is symmetric, then TTT \subseteq T^*. (For ψDT\psi \in \mathcal{D}_T, set η:=Tψ\eta := T\psi; symmetry gives ψTα=ηα\langle\psi|T\alpha\rangle = \langle\eta|\alpha\rangle for all αDT\alpha \in \mathcal{D}_T, so ψDT\psi \in \mathcal{D}_{T^*} with Tψ=TψT^*\psi = T\psi.) The whole game is to ask whether the inclusion is equality.

A maximality property#

Self-adjoint operators have no proper symmetric extensions. Concretely: if TT is self-adjoint, T~\tilde T is symmetric, and TT~T \subseteq \tilde T, then T=T~T = \tilde T.

Proof. From TT~T \subseteq \tilde T, the lemma gives T~T\tilde T^* \subseteq T^*. Symmetric means T~T~\tilde T \subseteq \tilde T^*, so T~T~T=T\tilde T \subseteq \tilde T^* \subseteq T^* = T. Combined with TT~T \subseteq \tilde T, we get T=T~T = \tilde T. \square

This is why self-adjointness is the right condition for observables. Spectral theorem guarantees self-adjoint operators have real spectra and a sensible functional calculus; symmetric operators in general do not. And the maximality property tells us that a self-adjoint operator is “as big as it can be” — there’s no room to extend it further while keeping the symmetric property.


3. Closability, closure, essentially self-adjoint#

There’s one more layer we need. Even if our operator TT isn’t self-adjoint, it may have a unique self-adjoint extension, hidden inside TT^*. This is captured by the notion of essentially self-adjointness.

Closures#

A densely defined operator TT is closable if its adjoint TT^* is itself densely defined. Equivalently (and more usefully), TT has a smallest closed extension, which we call the closure T\overline T. It satisfies T=T\overline T = T^{**}.

Symmetric operators are always closable. If TT is symmetric, then TTT \subseteq T^* implies DTDT\mathcal{D}_T \subseteq \mathcal{D}_{T^*}. Since DT\mathcal{D}_T is dense, so is DT\mathcal{D}_{T^*}, hence TT is closable. So we always have access to T=T\overline T = T^{**} for symmetric operators.

For symmetric TT, the closure sits between TT and TT^*:

TTT.T \subseteq T^{**} \subseteq T^*.

Essentially self-adjoint operators#

A symmetric operator TT is essentially self-adjoint (e.s.a.) if its closure T=T\overline T = T^{**} is self-adjoint.

This is weaker than self-adjointness. A self-adjoint operator is automatically e.s.a.: if T=TT = T^*, then taking adjoints gives T=TT^* = T^{**}, so T=TT = T^{**}, so T=T\overline T = T is self-adjoint.

Theorem. If TT is essentially self-adjoint, then T\overline T is the unique self-adjoint extension of TT.

Proof. T\overline T is a self-adjoint extension by hypothesis. For uniqueness, suppose SS is any other self-adjoint extension, TST \subseteq S. Self-adjoint operators are closed (since S=SS = S^* and adjoints are always closed), so TS\overline T \subseteq S (the closure is the smallest closed extension). But T\overline T is itself self-adjoint, and a self-adjoint operator has no proper symmetric extensions — so T=S\overline T = S. \square

The strategic upshot. When we want to elevate a symmetric operator to a genuine observable, we do not need it to already be self-adjoint. It suffices that it be essentially self-adjoint. Then the closure T\overline T is the canonical observable, picked out uniquely.

When essential self-adjointness fails, life gets harder: the symmetric operator may admit many self-adjoint extensions (or even none), and choosing one becomes a piece of physical input. The technology to count and parametrize these extensions is deficiency index theory, due to von Neumann — beyond our scope here, but coming up explicitly when we look at the interval case.


4. The momentum operator: setup#

Now to the main event. Define the momentum operator on the jj-th coordinate as

P^j:DPL2,ψijψ,\hat{P}_j: \mathcal{D}_P \to L^2,\qquad \psi \mapsto -i\,\partial_j\psi,

where the domain DP\mathcal{D}_P has yet to be specified. We work on H=L2([0,2π])\mathcal{H} = L^2([0, 2\pi]) in one dimension, and we’ll consider two natural-looking domains:

  • Interval (Dirichlet): DPint={ψC1([0,2π])ψ(0)=ψ(2π)=0}\mathcal{D}_P^{\text{int}} = \{\psi \in C^1([0, 2\pi]) \mid \psi(0) = \psi(2\pi) = 0\}.
  • Circle (periodic): DPcirc={ψC1([0,2π])ψ(0)=ψ(2π)}\mathcal{D}_P^{\text{circ}} = \{\psi \in C^1([0, 2\pi]) \mid \psi(0) = \psi(2\pi)\}.

The Dirichlet condition is strictly stronger than the periodic one (0=00 = 0 implies 0=00 = 0, but not conversely), so as operators we have P^jintP^jcirc\hat{P}_j^{\text{int}} \subsetneq \hat{P}_j^{\text{circ}} — the circle operator is an extension of the interval operator.

We’ll need a couple of function-space facts. The chain of inclusions

C1([a,b])H1([a,b])AC([a,b])C^1([a,b]) \subsetneq H^1([a,b]) \subseteq AC([a,b])

relates classically differentiable functions (C1C^1) to absolutely continuous functions (ACAC) and the Sobolev space H1H^1. Recall:

  • ψAC([a,b])\psi \in AC([a,b]) if there exists a Lebesgue-integrable ρ\rho such that ψ(x)=ψ(a)+axρ(y)dy\psi(x) = \psi(a) + \int_a^x \rho(y)\,dy, in which case ρ=ψ\rho = \psi' almost everywhere.
  • H1([a,b])={ψAC([a,b])ψL2}H^1([a,b]) = \{\psi \in AC([a,b]) \mid \psi' \in L^2\}.

The point of H1H^1 is that it’s the natural domain for ”ψ\psi has an L2L^2 derivative,” allowing us to apply ix-i\partial_x and stay in L2L^2without requiring classical differentiability everywhere.


5. Momentum on the interval [0,2π][0, 2\pi]#

Take P^jint\hat{P}_j^{\text{int}} with domain DP={ψC1([0,2π])ψ(0)=ψ(2π)=0}\mathcal{D}_P = \{\psi \in C^1([0, 2\pi]) \mid \psi(0) = \psi(2\pi) = 0\}.

Step 1. P^jint\hat{P}_j^{\text{int}} is symmetric#

For ψ,φDP\psi, \varphi \in \mathcal{D}_P, integrate by parts:

ψP^jφ=02πψ(x)(iφ(x))dx=i[ψφ]02π+i02πψ(x)φ(x)dx.\langle \psi | \hat{P}_j \varphi\rangle = \int_0^{2\pi} \overline{\psi(x)}\,\big(-i\varphi'(x)\big)\,dx = -i\big[\overline{\psi}\varphi\big]_0^{2\pi} + i \int_0^{2\pi} \overline{\psi'(x)}\,\varphi(x)\,dx.

The boundary term vanishes because ψ(0)=ψ(2π)=0\psi(0) = \psi(2\pi) = 0. The remaining integral is

i02πψφdx=02π(iψ)φdx=P^jψφ.i \int_0^{2\pi} \overline{\psi'}\,\varphi\,dx = \int_0^{2\pi} \overline{(-i\psi')}\,\varphi\,dx = \langle \hat{P}_j\psi | \varphi\rangle.

So P^jint\hat{P}_j^{\text{int}} is symmetric. ✓

Step 2. The adjoint (P^jint)(\hat{P}_j^{\text{int}})^* has a bigger domain#

Now we ask: for which ψL2\psi \in L^2 does there exist ηL2\eta \in L^2 with

02πψ(iφ)dx=02πηφdxfor all φDP?()\int_0^{2\pi} \overline{\psi}\,(-i\varphi')\,dx = \int_0^{2\pi} \overline\eta\,\varphi\,dx \qquad \text{for all } \varphi \in \mathcal{D}_P? \tag{$\star$}

The strategy: pick any NAC([0,2π])N \in AC([0,2\pi]) with N=ηN' = \eta a.e. (such an NN exists because every L2L^2 function on a compact interval has an absolutely continuous antiderivative). Substituting η=N\eta = N' and integrating the right-hand side by parts,

02πNφdx=[Nφ]02π02πNφdx,\int_0^{2\pi} \overline{N'}\,\varphi\,dx = \big[\overline{N}\varphi\big]_0^{2\pi} - \int_0^{2\pi} \overline N\,\varphi'\,dx,

and the boundary term vanishes because φDP\varphi \in \mathcal{D}_P vanishes at the endpoints. So ()(\star) becomes

02π(ψiN)φdx=0for all φDP.\int_0^{2\pi} \overline{(\psi - iN)}\,\varphi'\,dx = 0 \qquad \text{for all } \varphi \in \mathcal{D}_P.

Equivalently, ψiN{φφDP}\psi - iN \perp \{\varphi' \mid \varphi \in \mathcal{D}_P\} in L2L^2.

Now we identify this orthogonal complement. The set {φφDP}\{\varphi' \mid \varphi \in \mathcal{D}_P\} consists exactly of those continuous functions ξ\xi on [0,2π][0, 2\pi] satisfying 02πξ(x)dx=0\int_0^{2\pi} \xi(x)\,dx = 0 — that is, ξ1\xi \perp \mathbf{1}, where 1\mathbf{1} is the constant function. (One direction: if ξ=φ\xi = \varphi' with φ(0)=φ(2π)=0\varphi(0) = \varphi(2\pi) = 0, then 02πξdx=φ(2π)φ(0)=0\int_0^{2\pi}\xi\,dx = \varphi(2\pi) - \varphi(0) = 0. Conversely, given such a ξ\xi, the antiderivative φ(x)=0xξ(y)dy\varphi(x) = \int_0^x \xi(y)\,dy vanishes at both endpoints and lies in C1C^1.)

So the closure (in L2L^2) of {φφDP}\{\varphi' \mid \varphi \in \mathcal{D}_P\} is {1}\{\mathbf{1}\}^\perp, and its orthogonal complement is therefore {1}=span{1}=C1\{\mathbf{1}\}^{\perp\perp} = \overline{\operatorname{span}\{\mathbf{1}\}} = \mathbb{C}\cdot\mathbf{1}, the space of constant functions. Hence

ψiN=constantψ=constant+iN.\psi - iN = \text{constant} \quad\Longrightarrow\quad \psi = \text{constant} + iN.

Since NACN \in AC, this gives ψAC([0,2π])\psi \in AC([0, 2\pi]) with no boundary conditions whatsoever. The requirement P^jψ=η=iNL2\hat{P}_j^*\psi = \eta = -iN' \in L^2 then forces ψL2\psi' \in L^2, i.e. ψH1([0,2π])\psi \in H^1([0, 2\pi]). Conversely, every H1H^1 function arises this way, and so

 D(P^jint)=H1([0,2π]),(P^jint)ψ=iψ. \boxed{\ \mathcal{D}_{(\hat{P}_j^{\text{int}})^*} = H^1([0, 2\pi]),\qquad (\hat{P}_j^{\text{int}})^*\psi = -i\psi'.\ }

The adjoint is the same differential expression, but acting on a much larger domain — H1H^1 functions with no boundary conditions, versus C1C^1 functions vanishing at both endpoints.

In particular, P^jint(P^jint)\hat{P}_j^{\text{int}} \subsetneq (\hat{P}_j^{\text{int}})^*, so P^jint\hat{P}_j^{\text{int}} is not self-adjoint.

Step 3. The closure P^j\hat{P}_j^{**} is also too small#

Maybe the closure rescues us? Let’s compute DP\mathcal{D}_{P^{**}} directly.

ψDP\psi \in \mathcal{D}_{P^{**}} means: there exists ζL2\zeta \in L^2 such that ψP^jφ=ζφ\langle\psi | \hat P_j^* \varphi\rangle = \langle \zeta | \varphi\rangle for all φDP=H1\varphi \in \mathcal{D}_{P^*} = H^1. Since P^jP^j\hat P_j^{**} \subseteq \hat P_j^* and both act as ix-i\partial_x, ζ=iψ\zeta = -i\psi', and the question reduces to whether the appropriate boundary terms vanish.

For ψH1\psi \in H^1 (which we already know is necessary, since DPDP=H1\mathcal{D}_{P^{**}} \subseteq \mathcal{D}_{P^*} = H^1) and φH1\varphi \in H^1, integration by parts gives

ψP^jφP^jψφ=i[ψφ]02π.\langle\psi | \hat P_j^*\varphi\rangle - \langle \hat P_j^*\psi | \varphi\rangle = -i\big[\overline\psi\,\varphi\big]_0^{2\pi}.

Requiring this to vanish for all φH1\varphi \in H^1 — and H1H^1 functions can take arbitrary independent values at 00 and 2π2\pi — forces ψ(0)=0\psi(0) = 0 and ψ(2π)=0\psi(2\pi) = 0.

So

DP={ψH1([0,2π])ψ(0)=ψ(2π)=0},\mathcal{D}_{P^{**}} = \{\psi \in H^1([0, 2\pi]) \mid \psi(0) = \psi(2\pi) = 0\},

which is strictly smaller than DP=H1([0,2π])\mathcal{D}_{P^*} = H^1([0, 2\pi]). Hence P^jP^j\hat{P}_j^{**} \subsetneq \hat{P}_j^*, and P^jint\hat{P}_j^{\text{int}} is not essentially self-adjoint.

What’s really going on#

This is not a defect of our choice of “starting domain” — it reflects a real physical ambiguity. There is in fact a one-parameter family of self-adjoint extensions of P^jint\hat{P}_j^{\text{int}}, parametrized by θ[0,2π)\theta \in [0, 2\pi), with domains

Dθ={ψH1([0,2π])ψ(2π)=eiθψ(0)}.\mathcal{D}_\theta = \{\psi \in H^1([0, 2\pi]) \mid \psi(2\pi) = e^{i\theta}\psi(0)\}.

Different choices of θ\theta correspond to physically different theories (think: a particle on an interval threaded by a magnetic flux, or a “twisted” boundary condition). The deficiency-index analysis confirms that this is the full story: the deficiency indices of P^jint\hat P_j^{\text{int}} are (1,1)(1, 1), giving exactly a U(1)U(1)-worth of self-adjoint extensions.

The lesson: when an operator is symmetric but not essentially self-adjoint, the physics of which observable to use is not determined by the formal expression alone. You have to add a choice of boundary condition, and that choice is part of defining the system.


6. Momentum on the circle#

Now repeat the analysis with the milder boundary condition ψ(0)=ψ(2π)\psi(0) = \psi(2\pi) (no requirement that the common value be zero).

Step 1. Symmetric#

For ψ,φ\psi, \varphi with ψ(0)=ψ(2π)\psi(0) = \psi(2\pi) and φ(0)=φ(2π)\varphi(0) = \varphi(2\pi), the boundary term in integration by parts is

i[ψφ]02π=i(ψ(2π)φ(2π)ψ(0)φ(0))=iψ(0)(φ(2π)φ(0))=0,-i\big[\overline\psi\varphi\big]_0^{2\pi} = -i\big(\overline{\psi(2\pi)}\varphi(2\pi) - \overline{\psi(0)}\varphi(0)\big) = -i\overline{\psi(0)}\big(\varphi(2\pi) - \varphi(0)\big) = 0,

using both boundary conditions. So P^jcirc\hat P_j^{\text{circ}} is symmetric. ✓

Step 2. The adjoint#

We can shortcut the computation by using the inclusion P^jintP^jcirc\hat{P}_j^{\text{int}} \subsetneq \hat{P}_j^{\text{circ}}. By the inclusion lemma, (P^jcirc)(P^jint)=P^jH1(\hat{P}_j^{\text{circ}})^* \subseteq (\hat{P}_j^{\text{int}})^* = \hat{P}_j^*|_{H^1}. So D(Pcirc)H1([0,2π])\mathcal{D}_{(P^{\text{circ}})^*} \subseteq H^1([0, 2\pi]) and the action is still ix-i\partial_x.

To pin down the boundary conditions: ψD(Pcirc)\psi \in \mathcal{D}_{(P^{\text{circ}})^*} requires the boundary term to vanish for all φ\varphi with φ(0)=φ(2π)\varphi(0) = \varphi(2\pi):

i[ψφ]02π=i(ψ(2π)ψ(0))φ(0)=0for all such φ.-i\big[\overline\psi\varphi\big]_0^{2\pi} = -i\overline{\big(\psi(2\pi) - \psi(0)\big)}\,\varphi(0) = 0 \qquad \text{for all such } \varphi.

Since φ(0)\varphi(0) is arbitrary, this forces ψ(2π)=ψ(0)\psi(2\pi) = \psi(0). So

 D(P^jcirc)={ψH1([0,2π])ψ(0)=ψ(2π)}. \boxed{\ \mathcal{D}_{(\hat{P}_j^{\text{circ}})^*} = \{\psi \in H^1([0, 2\pi]) \mid \psi(0) = \psi(2\pi)\}.\ }

The adjoint inherits the same periodic boundary condition, just on a bigger function space.

Step 3. Not self-adjoint, but…#

Comparing,

DPcirc={ψC1ψ(0)=ψ(2π)}{ψH1ψ(0)=ψ(2π)}=D(Pcirc).\mathcal{D}_{P^{\text{circ}}} = \{\psi \in C^1 \mid \psi(0) = \psi(2\pi)\} \subsetneq \{\psi \in H^1 \mid \psi(0) = \psi(2\pi)\} = \mathcal{D}_{(P^{\text{circ}})^*}.

So P^jcirc\hat{P}_j^{\text{circ}} is not self-adjoint either — but only because of the regularity gap C1H1C^1 \subsetneq H^1, not because of any boundary-condition mismatch.

Step 4. Essentially self-adjoint!#

Let’s compute the closure. For ψDP\psi \in \mathcal{D}_{P^{**}} and φDP={H1 with periodic BC}\varphi \in \mathcal{D}_{P^*} = \{H^1\text{ with periodic BC}\}, the boundary term is

i[ψφ]02π=i(ψ(2π)ψ(0))φ(0)-i\big[\overline\psi\varphi\big]_0^{2\pi} = -i\overline{\big(\psi(2\pi) - \psi(0)\big)}\,\varphi(0)

(using φ(2π)=φ(0)\varphi(2\pi) = \varphi(0)). For this to vanish for all such φ\varphi, we need ψ(0)=ψ(2π)\psi(0) = \psi(2\pi) — but no further constraint, because φ(0)\varphi(0) is the only free quantity. So

DP={ψH1([0,2π])ψ(0)=ψ(2π)}=DP.\mathcal{D}_{P^{**}} = \{\psi \in H^1([0, 2\pi]) \mid \psi(0) = \psi(2\pi)\} = \mathcal{D}_{P^*}.

Therefore P^j=P^j\hat{P}_j^{**} = \hat{P}_j^*, and P^jcirc\hat{P}_j^{\text{circ}} is essentially self-adjoint.

Conclusion#

The momentum operator on the circle has a unique self-adjoint realization: take the closure of the naive operator on C1C^1-with-periodic-BC, and you land on the operator ix-i\partial_x with domain {ψH1ψ(0)=ψ(2π)}\{\psi \in H^1 \mid \psi(0) = \psi(2\pi)\}.

This is the operator whose eigenfunctions are einxe^{inx} with eigenvalues nZn \in \mathbb{Z}, which is exactly the spectrum of momentum on S1S^1 that you’d compute formally in any QM textbook. The point of the whole apparatus is to justify that calculation by showing that the operator we’re diagonalizing actually exists as a genuine observable.


7. The takeaway#

SettingSymmetric?Self-adjoint?Essentially self-adjoint?Self-adjoint extensions
Interval [0,2π][0,2\pi], ψ(0)=ψ(2π)=0\psi(0) = \psi(2\pi) = 0yesnonoone-parameter family θU(1)\theta \in U(1)
Circle, ψ(0)=ψ(2π)\psi(0) = \psi(2\pi)yesno (regularity only)yesunique: the closure

The interval case is a subtle example of a phenomenon that recurs throughout quantum mechanics: a perfectly innocent-looking differential operator may fail to define a unique observable, because the boundary conditions encode physical information that cannot be recovered from the formal expression alone. The circle case is the happier scenario where everything works out, and it works out because the topology of the configuration space already does the boundary-condition-fixing for us.

The general technology to handle the harder cases — counting the deficiency indices (n+,n)(n_+, n_-) to find a U(n+)U(n_+)-family of self-adjoint extensions when n+=nn_+ = n_-, or no extensions at all when n+nn_+ \neq n_- — is due to von Neumann. It’s the right machinery for asking systematically “when is this symmetric operator a genuine observable, and if not, what data is missing?”

We also haven’t said anything about why momentum is ix-i\partial_x in the first place. That’s the content of the Stone–von Neumann theorem, which from the canonical commutation relation [x^,p^]=i[\hat x, \hat p] = i (and an irreducibility hypothesis) recovers the Schrödinger representation up to unitary equivalence. The classical-mechanics analogue is the Poisson bracket {x,p}=1\{x, p\} = 1; quantization replaces Poisson brackets by commutators, and Stone–von Neumann tells you that this prescription has, up to mathematical niceness conditions, essentially one realization. That’s a separate story for another post.


References#

  • B. C. Hall, Quantum Theory for Mathematicians, Springer (2013) — Chapters 7–10.
  • M. Reed and B. Simon, Methods of Modern Mathematical Physics, Volume I: Functional Analysis, Academic Press (1980) — especially Chapters VI–VIII.
  • F. Schuller, Lectures on Quantum Theory, University of Erlangen-Nürnberg, YouTube playlist (Lectures 6–9 cover this material).
  • P. Szekeres, A Course in Modern Mathematical Physics: Groups, Hilbert Space and Differential Geometry, Cambridge University Press (2004) — Chapters 13–14 for background reading.
When 'Hermitian' Isn't Enough: A Case Study of the Momentum Operator
https://rohankulkarni.me/posts/notes/case-study-momentum/
Author
Rohan Kulkarni
Published at
2021-12-01
License
CC BY-NC-SA 4.0
Comments