English Reciprocals and the Limits of Categorization
Brett Reynolds
Humber Polytechnic & University of Toronto
2026-04-10
The puzzle
Each other and one another
Everyone calls them pronouns.
But how do we know that’s the right lexical category?
And when there’s only two of them, …?
What makes a category real?
Traditional answer: necessary and sufficient conditions
Problem: linguistic categories resist definition. Pronouns share properties, but no single property is shared by all and only pronouns.
Alternative: a robust cluster of co-occurring properties + stabilizing mechanisms that maintain the cluster = a real category
Homeostatic = self-correcting, like a Watt governor: the system drifts, mechanisms push it back. A category maintained this way is projectible – you can make predictions about new instances.
No real kind without a purpose
Syntactician’s proper noun
Distribution: bare NP position
Agreement: 3rd person singular
Modification: restricted
- Typically has proper name semantics
Semanticist’s proper name
Rigid designation
Referential opacity
Sense vs. reference
- Usually instantiated by a proper noun
Brett is both. Different projections, different mechanisms, different HPCs, same extension.
Homeostasis: the virtuous circle
What holds these clusters together?
Property cluster co-occurring properties
sustains →
← stabilizes
Mechanisms causal processes
Mechanisms maintaining grammatical categories:
Acquisition – children converge on categories from distributional input
Entrenchment – high-frequency items anchor the cluster
Interactive alignment – speakers converge in conversation
Iterated transmission – learnable structure survives across generations
Functional pressure – categories persist because they’re useful
Mechanisms maintain clusters. Clusters maintain mechanisms. That’s what homeostatic means. (A reciprocal relationship, as it happens.)
Stability is dynamic, not static
Grammatical categories are spinning tops, not balls in valleys.
The data
I gathered every property I could think of, however trivial, and coded them for all the CGEL pronouns (65) and determinatives (73).
Property
each other
one another
they
somebody
Monomorphemic
Y
Definite
Y
Y
Y
Anaphoric
Y
Y
Y
Y
Fused determiner-head
Y
Appears in object
Y
Y
Y
Requires antecedent
Y
Y
155 binary properties × 138 items. The goal: leave no room for cherry-picking.
The reciprocals puzzle
Pronoun-like
Determinative-like
Morphology (66)
Compound; no distinct accusative, genitive, or reflexive forms
Semantics (36)
Definite; anaphoric; requires an antecedent
Syntax (50)
Not in partitives; not in existentials; no else
Accepts ’s; appears in object
Phonology (3)
(weak signal)
Morphology pulls one way, semantics the other, syntax is mixed. Which way do they go?
The problem with cherry-picking
Two items, 155 tests, and a strong temptation to cherry-pick.
Croft (2001) calls this methodological opportunism: consciously or not, we select tests that support our preferred analysis.
The alternative: measure the stability of diagnostic ambiguity. Vary every reasonable analytic choice and ask whether the answer changes.
The interesting question isn’t “which category?” but “how stable is the apparent boundary position under different measurement choices?”
What HPC predicts for boundary items
Stable position: the result doesn’t depend on how you measure
Cross-dimensional tension: morphology and semantics pull in different directions
Clean anchors: clear cases come out right, so the method is trustworthy
Near-parity mixture: the item sits right at the midpoint between the two categories
Robustness to null: scramble the data keeping its basic structure; the pattern shouldn’t appear by chance — and it doesn’t
These aren’t arbitrary desiderata. They’re consequences of the theory.
Mapping grammatical space
155 binary properties (morphology, syntax, semantics, phonology) across 138 items. This 2D projection captures ~17% of the variance; all actual measurement uses full 155-dimensional Jaccard distances.
Multiple Correspondence Analysis projection. Pronouns (blue) and determinatives (red) form regions; compound determinatives sit at the interface; reciprocals (triangles) fall in that interface zone.
Not a statistical fluke
Scramble the data 5,000 times, preserving how many properties each word has and how many words have each property. This tests whether the specific combination of features drives reciprocals’ position, not just marginal structure.
Observed pattern in only 0.6% of scrambles (p = 0.006).
Permutation null distribution. Dashed line marks the observed value.
Stable across analytic choices
Vary every reasonable analytic choice (distance metric, which properties, weighting) and show all results. Each point is one specification; Delta = mean distance to pronouns minus mean distance to determinatives.
Each point is one analytic specification: different distance metrics, different feature weightings. Positive = closer to determinatives; negative = closer to pronouns.
Sign stable across most choices. Removing morphology flips it. That’s cross-dimensional tension.
Right at the midpoint
Best-fitting mixture weight: each other ~0.5, one another ~0.5. Remove morphology: both jump to ~0.94 (strongly pronoun-like). Remove semantics: both shift toward determinative (Delta ~ +0.09). The midpoint exists because morphology and semantics are pulling in opposite directions.
Every item sorted from determinative (0) to pronoun (1). Reciprocals sit at the midpoint.
All five expectations confirmed
Expectation
Result
✔
Stable position
Result stable no matter how you measure
✔
Cross-dimensional tension
Morphology → determinative; semantics → pronoun
✔
Clean anchors
Same methods correctly identify clear cases
✔
Near-parity mixture
Best-fitting weights ~0.53, ~0.49 (near midpoint)
✔
Robustness to null
Pattern in only 0.6% of scrambled data
This isn’t measurement failure. It’s what a real boundary looks like – and it tells you what you can predict: roughly half a pronoun’s behaviour, half a determinative’s.
What kind of problem is this?
Reciprocals are one or the other. But our instruments can’t resolve which.
ResolvedUnresolved
Categories are internally gradient but sharply bounded. This isn’t gradience; it’s a boundary phenomenon: independent mechanisms sustaining opposed pulls.
Test against scrambled baselines (especially with small n)
Vary specifications systematically (show all results)
Calibrate against clear cases (verify known structure)
Ask whether the ambiguity is stable
Cash out the projective consequences (what does the classification predict?)
Categories are real because they’re projectible. Maintenance is the mechanism; projection is the payoff. Stable ambiguity tells you exactly how much projection each anchor category provides.