How to Study Boundary Phenomena

English Reciprocals and the Limits of Categorization

Brett Reynolds

Humber Polytechnic & University of Toronto

2026-04-10

The puzzle

Each other and one another


Everyone calls them pronouns.


But how do we know that’s the right lexical category?


And when there’s only two of them, …?

What makes a category real?

  • Traditional answer: necessary and sufficient conditions
  • Problem: linguistic categories resist definition. Pronouns share properties, but no single property is shared by all and only pronouns.
  • Alternative: a robust cluster of co-occurring properties + stabilizing mechanisms that maintain the cluster = a real category

Homeostatic = self-correcting, like a Watt governor: the system drifts, mechanisms push it back. A category maintained this way is projectible – you can make predictions about new instances.

No real kind without a purpose

Syntactician’s proper noun

  • Distribution: bare NP position
  • Agreement: 3rd person singular
  • Modification: restricted

- Typically has proper name semantics

Semanticist’s proper name

  • Rigid designation
  • Referential opacity
  • Sense vs. reference

- Usually instantiated by a proper noun

Brett is both. Different projections, different mechanisms, different HPCs, same extension.

Homeostasis: the virtuous circle

What holds these clusters together?

Property cluster
co-occurring properties

sustains →

← stabilizes

Mechanisms
causal processes

Mechanisms maintaining grammatical categories:

  1. Acquisition – children converge on categories from distributional input
  2. Entrenchment – high-frequency items anchor the cluster
  3. Interactive alignment – speakers converge in conversation
  4. Iterated transmission – learnable structure survives across generations
  5. Functional pressure – categories persist because they’re useful


Mechanisms maintain clusters. Clusters maintain mechanisms. That’s what homeostatic means. (A reciprocal relationship, as it happens.)

Stability is dynamic, not static

Grammatical categories are spinning tops, not balls in valleys.

The data

I gathered every property I could think of, however trivial, and coded them for all the CGEL pronouns (65) and determinatives (73).

Property each other one another they somebody
Monomorphemic Y
Definite Y Y Y
Anaphoric Y Y Y Y
Fused determiner-head Y
Appears in object Y Y Y
Requires antecedent Y Y

155 binary properties × 138 items. The goal: leave no room for cherry-picking.

The reciprocals puzzle

Pronoun-like Determinative-like
Morphology (66) Compound; no distinct accusative, genitive, or reflexive forms
Semantics (36) Definite; anaphoric; requires an antecedent
Syntax (50) Not in partitives; not in existentials; no else Accepts ’s; appears in object
Phonology (3) (weak signal)

Morphology pulls one way, semantics the other, syntax is mixed. Which way do they go?

The problem with cherry-picking

Two items, 155 tests, and a strong temptation to cherry-pick.

Croft (2001) calls this methodological opportunism: consciously or not, we select tests that support our preferred analysis.

The alternative: measure the stability of diagnostic ambiguity. Vary every reasonable analytic choice and ask whether the answer changes.

The interesting question isn’t “which category?” but “how stable is the apparent boundary position under different measurement choices?”

What HPC predicts for boundary items

  • Stable position: the result doesn’t depend on how you measure
  • Cross-dimensional tension: morphology and semantics pull in different directions
  • Clean anchors: clear cases come out right, so the method is trustworthy
  • Near-parity mixture: the item sits right at the midpoint between the two categories
  • Robustness to null: scramble the data keeping its basic structure; the pattern shouldn’t appear by chance — and it doesn’t

These aren’t arbitrary desiderata. They’re consequences of the theory.

Mapping grammatical space

155 binary properties (morphology, syntax, semantics, phonology) across 138 items. This 2D projection captures ~17% of the variance; all actual measurement uses full 155-dimensional Jaccard distances.

Multiple Correspondence Analysis projection. Pronouns (blue) and determinatives (red) form regions; compound determinatives sit at the interface; reciprocals (triangles) fall in that interface zone.

Not a statistical fluke

Scramble the data 5,000 times, preserving how many properties each word has and how many words have each property. This tests whether the specific combination of features drives reciprocals’ position, not just marginal structure.

Observed pattern in only 0.6% of scrambles (p = 0.006).

Permutation null distribution. Dashed line marks the observed value.

Stable across analytic choices

Vary every reasonable analytic choice (distance metric, which properties, weighting) and show all results. Each point is one specification; Delta = mean distance to pronouns minus mean distance to determinatives.

Each point is one analytic specification: different distance metrics, different feature weightings. Positive = closer to determinatives; negative = closer to pronouns.

Sign stable across most choices. Removing morphology flips it. That’s cross-dimensional tension.

Right at the midpoint

Best-fitting mixture weight: each other ~0.5, one another ~0.5. Remove morphology: both jump to ~0.94 (strongly pronoun-like). Remove semantics: both shift toward determinative (Delta ~ +0.09). The midpoint exists because morphology and semantics are pulling in opposite directions.

Every item sorted from determinative (0) to pronoun (1). Reciprocals sit at the midpoint.

All five expectations confirmed

Expectation Result
Stable position Result stable no matter how you measure
Cross-dimensional tension Morphology → determinative; semantics → pronoun
Clean anchors Same methods correctly identify clear cases
Near-parity mixture Best-fitting weights ~0.53, ~0.49 (near midpoint)
Robustness to null Pattern in only 0.6% of scrambled data


This isn’t measurement failure. It’s what a real boundary looks like – and it tells you what you can predict: roughly half a pronoun’s behaviour, half a determinative’s.

What kind of problem is this?

Reciprocals are one or the other. But our instruments can’t resolve which.

Resolved Unresolved

Categories are internally gradient but sharply bounded. This isn’t gradience; it’s a boundary phenomenon: independent mechanisms sustaining opposed pulls.

How to study boundary phenomena

  1. Build comprehensive profiles (don’t cherry-pick diagnostics)
  2. Test against scrambled baselines (especially with small n)
  3. Vary specifications systematically (show all results)
  4. Calibrate against clear cases (verify known structure)
  5. Ask whether the ambiguity is stable
  6. Cash out the projective consequences (what does the classification predict?)


Categories are real because they’re projectible. Maintenance is the mechanism; projection is the payoff. Stable ambiguity tells you exactly how much projection each anchor category provides.


Paper: LingBuzz 009294 · Code: GitHub · R package in progress · brettreynolds.ca · brett.reynolds@humber.ca