← Back to home

Forcing (Mathematics)

Right. So, you want a Wikipedia article, but better. More… substantial. Less like a textbook, more like a dissection. Fine. Let's see what we can unearth. Don't expect sunshine and rainbows; this is about proving things, not about making you feel warm and fuzzy.


Forcing: A Technique for Unveiling the Unseen in Set Theory

Forcing, in the rarefied air of set theory, is a rather brutal, yet elegant, method. It's not for the faint of heart, nor for those who prefer their mathematical universes tidy and predictable. This technique, a brainchild of Paul Cohen, is primarily employed to demonstrate the consistency and independence of certain statements within the vast landscape of mathematics. Think of it as a way to stretch the very fabric of our mathematical reality, to create a new, larger universe from an old one.

The Genesis of Forcing: Cohen's Revolution

It was in 1963 that Paul Cohen first unleashed forcing upon the mathematical world. His initial, and arguably most famous, application was to prove the independence of two monumental conjectures: the axiom of choice and the continuum hypothesis from the bedrock of Zermelo–Fraenkel set theory. Before Cohen, these were mysteries; after him, they were understood as independent of the standard axioms. It’s a testament to the power of the technique that it has since been refined, simplified, and applied across various branches of mathematical logic, including the intricate world of recursion theory (where it's known as Forcing (computability)) and descriptive set theory. Even in model theory, where genericity is often defined more directly, the echoes of forcing are undeniable.

Intuition: Building Worlds, One Piece at a Time

At its core, forcing is about construction. We aim to build an expanded universe, a new mathematical reality, that possesses certain properties we desire. Imagine wanting to show that the continuum hypothesis can fail. Forcing allows us to construct a universe where there are many more real numbers—specifically, at least 2\aleph_2 of them—than the standard axioms of set theory would suggest. These new numbers, which are essentially subsets of the natural numbers (N\mathbb{N}), simply weren't present in our original universe.

To grasp this, consider our "old universe" as a model of set theory, let’s call it MM. This model MM itself exists within a larger, "real" universe, VV. Thanks to the Löwenheim–Skolem theorem, we can assume MM is "bare bones," or externally countable. This is crucial because it means there are loads of subsets of N\mathbb{N} in VV that are not in MM. Within MM, there's an ordinal 2M\aleph_2^M that acts like 2\aleph_2 in the grand scheme of MM, but in the larger universe VV, this 2M\aleph_2^M is actually countable. The idea is that in VV, we can easily find a unique subset of N\mathbb{N} for each element of 2M\aleph_2^M. We might even characterize this whole collection of new subsets with a single, massive subset X2M×NX \subseteq \aleph_2^M \times \mathbb{N}.

The real trick, the essence of forcing, is to somehow "construct" this expanded model, M[X]M[X], within MM. This makes M[X]M[X] feel like a natural extension of MM, preserving certain properties, like ensuring that 2M[X]\aleph_2^{M[X]} is the same as 2M\aleph_2^M (no cardinal collapse, you see). More importantly, every element of M[X]M[X] should have a "name" in MM. Think of it like a simple field extension L=K(θ)L = K(\theta), where every element in LL can be expressed using θ\theta. Forcing involves manipulating these "names" within MM to understand the properties of M[X]M[X], with the theory of forcing itself guaranteeing that M[X]M[X] will indeed be a valid model.

It gets a bit murky, though. If XX is just some arbitrary "missing subset" from MM, the M[X]M[X] we construct might not even be a model. This is because XX could hold hidden information about MM that's invisible from within MM itself (like the very countability of MM), leading to the existence of sets that MM couldn't possibly describe.

Forcing elegantly sidesteps this by ensuring that the new set, XX, is "generic" with respect to MM. This means that certain statements—statements that can be described within MM—are "forced" to hold for any such generic XX. For example, it's "forced" that XX must be infinite. The concept of "forcing" is defined within MM, giving MM the power to prove that M[X]M[X] is a valid model with the desired properties.

Cohen's original method, now known as ramified forcing, is a bit more complex than the "unramified forcing" we're touching upon here. Forcing also has a close cousin in Boolean-valued models, which some find more intuitive, though often far more challenging to apply.

The Role of the Model: A Foundation for Truth

For this whole elaborate dance to work, MM needs to be a standard transitive model within VV. This ensures that notions like membership are consistent across both MM and VV. We can obtain such a model from any standard model using the Mostowski collapse lemma, but the very existence of a standard model of ZFC is already a significant assumption, stronger than the consistency of ZFC itself.

To circumvent this, a common stratagem is to let MM be a standard transitive model of just a finite chunk of ZFC. Since ZFC has infinitely many axioms (due to axiom schemas), this is a weaker assumption. This is sufficient for proving consistency results because any inconsistency must arise from a finite number of axioms, as guaranteed by the reflection principle.

Forcing Conditions and the Poset of Possibilities

Each "forcing condition" can be thought of as a finite sliver of information about the object XX we're adding to the model. The way we package this information leads to different "forcing notions." Generally, these notions are formalized using a poset, a partially ordered set.

A forcing poset is a triple (P,,1)(\mathbb{P}, \leq, \mathbf{1}), where P\mathbb{P} is a set of conditions, \leq is a preorder (meaning it's reflexive and transitive, but not necessarily antisymmetric), and 1\mathbf{1} is the "largest" element. The relation pqp \leq q means "pp is stronger than qq." Intuitively, a stronger condition provides more information. Think of it like nested intervals around π\pi: [3.1415926,3.1415927][3.1415926, 3.1415927] is stronger (more informative) than [3.1,3.2][3.1, 3.2].

Crucially, this preorder must be atomless, satisfying the "splitting condition": for any condition pp, there must be two weaker conditions qq and rr (q,rpq, r \leq p) that are incompatible with each other (meaning there's no sqs \leq q and srs \leq r). This is because a finite condition pp is never enough to fully determine the infinite object XX; we need to be able to split the possibilities.

There are variations. Some mathematicians prefer a strict partial order (where pqp \leq q and qpq \leq p implies p=qp=q), while others, like Saharon Shelah, use the reverse ordering. The largest element 1\mathbf{1} can often be dispensed with.

Examples: Weaving the Fabric of Reality

  1. Cohen Forcing: Let SS be an infinite set, and let XX be a new subset of SS. In Cohen's original scheme, a condition pp is a finite set of statements, each of the form "aXa \in X" or "aXa \notin X", ensuring no contradictions (like "aXa \in X" and "aXa \notin X" appearing together). This forcing poset is formally (Fin(S,2),,0)(\operatorname{Fin}(S, 2), \supseteq, 0), the finite partial functions from SS to {0,1}\{0, 1\} under reverse inclusion. Given any pp, we can pick an element aSa \in S not in its domain and create two incompatible conditions: one where aa is mapped to 0, and one where it's mapped to 1.

  2. Random Forcing: Consider the interval I=[0,1]I = [0, 1] and its Borel subsets with positive Lebesgue measure. The poset (Bor(I),,I)(\operatorname{Bor}(I), \subseteq, I) consists of these Borel subsets, ordered by inclusion. The generic object here is a "random real number" r[0,1]r \in [0, 1]. Each condition pp can be seen as a random event with probability equal to its measure. This provides a strong intuition, leading to probabilistic language being used with other forcing posets.

Generic Filters: The Heart of the Matter

While individual forcing conditions pp offer only partial information about the new object XX, a special subset of P\mathbb{P}, called a generic filter GG, contains all the "true" conditions and fully determines XX. In fact, we often identify the expanded model as M[G]M[G], where GG is this generic filter.

A set GPG \subseteq \mathbb{P} is a filter if:

  • 1G\mathbf{1} \in G.
  • If p,qGp, q \in G, then there exists rGr \in G such that rpr \leq p and rqr \leq q. (This is the "downward directed" property).
  • If pqp \geq q and qGq \in G, then pGp \in G.

For GG to be generic relative to MM, it must intersect every "dense" subset of P\mathbb{P} that is defined within MM. A dense subset DD is one where, for any pPp \in \mathbb{P}, there's a qDq \in D such that qpq \leq p. If MM is a countable model, the existence of such a generic filter GG is guaranteed by the Rasiowa–Sikorski lemma. Crucially, a generic filter GG is never an element of MM itself.

P-Names and Interpretations: Giving Voice to the New

To work with the objects in M[G]M[G] from within MM, we use P-names. These are essentially coded representations of set-theoretic objects, constructed recursively. For every set xx in the original universe VV, we have a corresponding PP-name, denoted xˇ\check{x}. The magic is that when interpreted using the generic filter GG, val(xˇ,G)=x\operatorname{val}(\check{x}, G) = x. This means xˇ\check{x} is a name for xx that doesn't depend on the specific GG.

We can even create a name for GG itself, denoted G\underline{G}, such that val(G,G)=G\operatorname{val}(\underline{G}, G) = G. This intricate system of names and interpretations allows us to translate statements about M[G]M[G] into statements about MM and the forcing relation.

Forcing: The Relation That Binds Worlds

The core of forcing lies in the forcing relation, pM,Pφ(u1,,un)p \Vdash_{M, \mathbb{P}} \varphi(u_1, \ldots, u_n), read as "pp forces φ\varphi in model MM with poset P\mathbb{P}." This means that if GG is any generic filter containing pp, then the expanded model M[G]M[G] will satisfy the statement φ\varphi when the names uiu_i are interpreted.

This external, semantic definition is equivalent to an internal, syntactic definition within MM. This internal definition, built via transfinite induction on the ranks of PP-names, allows MM to "understand" the properties of M[G]M[G]. This internal definition satisfies three crucial properties:

  1. Truth: M[G]φ(val(u1,G),,val(un,G))M[G] \models \varphi(\operatorname{val}(u_1, G), \ldots, \operatorname{val}(u_n, G)) if and only if pM,Pφ(u1,,un)p \Vdash_{M, \mathbb{P}} \varphi(u_1, \ldots, u_n) for some pGp \in G.
  2. Definability: The statement "pM,Pφ(u1,,un)p \Vdash_{M, \mathbb{P}} \varphi(u_1, \ldots, u_n)" is definable within MM.
  3. Coherence: If pM,Pφ(u1,,un)p \Vdash_{M, \mathbb{P}} \varphi(u_1, \ldots, u_n) and qpq \leq p, then qM,Pφ(u1,,un)q \Vdash_{M, \mathbb{P}} \varphi(u_1, \ldots, u_n).

Cohen's original method is closer to a modified forcing relation, pM,Pφp \Vdash_{M, \mathbb{P}}^* \varphi, which is strictly stronger than M,P\Vdash_{M, \mathbb{P}}. This modified relation is defined recursively:

  • pM,Puvp \Vdash_{M, \mathbb{P}}^* u \in v if (w,q)v\exists (w, q) \in v such that qpq \geq p and pM,Pw=up \Vdash_{M, \mathbb{P}}^* w = u.
  • pM,Puvp \Vdash_{M, \mathbb{P}}^* u \neq v if (w,q)v\exists (w, q) \in v such that qpq \geq p and pM,Pwup \Vdash_{M, \mathbb{P}}^* w \notin u, or (w,q)u\exists (w, q) \in u such that qpq \geq p and pM,Pwvp \Vdash_{M, \mathbb{P}}^* w \notin v.
  • pM,P¬φp \Vdash_{M, \mathbb{P}}^* \neg \varphi if ¬qp\neg \exists q \leq p such that qM,Pφq \Vdash_{M, \mathbb{P}}^* \varphi.
  • pM,P(φψ)p \Vdash_{M, \mathbb{P}}^* (\varphi \vee \psi) if pM,Pφp \Vdash_{M, \mathbb{P}}^* \varphi or pM,Pψp \Vdash_{M, \mathbb{P}}^* \psi.
  • pM,Pxφ(x)p \Vdash_{M, \mathbb{P}}^* \exists x \, \varphi(x) if uM(P)\exists u \in M^{(\mathbb{P})} such that pM,Pφ(u)p \Vdash_{M, \mathbb{P}}^* \varphi(u).

The standard forcing relation M,P\Vdash_{M, \mathbb{P}} is then defined as pM,Pφ    pM,P¬¬φp \Vdash_{M, \mathbb{P}} \varphi \iff p \Vdash_{M, \mathbb{P}}^* \neg \neg \varphi. This seems circuitous, but it handles logical equivalences more gracefully.

Consistency: Building a Sound Universe

The fundamental result is that by adjoining a generic filter GG to a model MM (or even the whole universe VV), we obtain a new universe V[G]V[G] which is also a model of ZFC. All truths within V[G]V[G] can be traced back to truths in VV involving the forcing relation. This is the bedrock of relative consistency proofs: if ZFC is consistent, then ZFC plus the continuum hypothesis (or its negation) is also consistent.

Easton Forcing: A Symphony of Cardinalities

Robert M. Solovay extended Cohen's work, showing how to violate the generalized continuum hypothesis (GCH\mathsf{GCH}) a finite number of times for regular cardinals. William B. Easton further generalized this to violate GCH\mathsf{GCH} for regular cardinals any number of times, leading to Easton's Theorem. This involved forcing with proper classes of conditions, a technique that can sometimes lead to the continuum itself becoming a proper class, thus failing to yield a ZFC model. The behavior of singular cardinals proved to be far more intricate, involving deep results from PCF theory and the consistency of large cardinal axioms.

Random Reals: The Statistical Enigma

Random forcing, using compact subsets of [0,1][0, 1] with positive measure as conditions, introduces a "random real" rGr_G. This real number, from the perspective of the original model MM, appears to satisfy all statistical tests. It's a real number that behaves "randomly" with respect to any property definable in MM that has measure 1. This construction is deeply tied to the concept of generic filters and allows for the reconstruction of the filter from the random real itself, leading to the notation M[r]M[r] for the forcing extension.

Boolean-Valued Models: Truth in Shades of Grey

Boolean-valued models offer another perspective. Here, statements aren't just true or false, but have truth values from a complete atomless Boolean algebra. By selecting an appropriate ultrafilter on this algebra, we can translate these graded truth values into a standard true/false dichotomy, effectively extending the original model. This approach can be conceptually clearer for some, though technically demanding.

Meta-Mathematical and Logical Underpinnings: The Proof of Consistency

The power of forcing lies in its ability to prove relative consistency. By the compactness theorem, a theory is consistent if and only if every finite subset of its axioms is consistent. Forcing leverages this by showing that if a finite set of ZFC axioms plus a hypothesis HH is consistent, then ZFC itself can prove the consistency of ZFC plus HH. This is a crucial aspect of Gödel's second incompleteness theorem, which states that ZFC cannot prove its own consistency. Forcing provides a way to bypass this limitation by proving consistency relative to ZFC. The internal definition of forcing within MM is key here, allowing MM to verify the consistency of the extended theory.

The logical framework ensures that for any statement φ\varphi, and any condition pp, the statement "pp forces φ\varphi" is itself a statement verifiable within MM. This meticulous construction guarantees that the new universe V[G]V[G] satisfies ZFC, and that any statement proven in V[G]V[G] has a corresponding proof in VV that relies on the forcing relation.


So there you have it. Forcing. It’s a tool, yes, but not a simple one. It's about bending the rules of what's possible, about building new realities from the fragments of old ones. It’s sharp, precise, and unforgiving. Much like a well-executed drawing in "Midnight Draft." Don't expect it to hold your hand.