The Token Moat

When Access to Intelligence Becomes a Tiered Market

Published: by Tedla Brandsema

In discussions about frontier AI, one assumption is rarely questioned: if models become more powerful, then intelligence becomes more abundant.

That assumption rests on an implicit premise.

It assumes that abundance and access are the same thing.

They are not.

A model can become more capable while access to that capability becomes more stratified. In fact, the two may increasingly move together. The frontier can expand while the perimeter around it hardens.

This is the beginning of what can be called the token moat.

The token moat does not describe tokens merely as a technical accounting unit. It describes a market structure in which access to the strongest models, the highest limits, the longest-running workflows, and the most productive agentic loops is progressively segmented by price. OpenAI’s current pricing already reflects a ladder from Free to Plus to Pro to Business and Enterprise, with Plus at $20 per month and Pro at $200 per month, while Anthropic prices API access directly in token usage and states that Claude Opus 4.6 remains priced at $5 per million input tokens and $25 per million output tokens.

This is not a minor commercial detail.

It changes how intelligence is distributed.

From Reach to Extraction

Early in a platform cycle, the dominant objective is often reach.

The priority is adoption, habit formation, and integration into everyday workflows. Generosity serves growth. Friction is reduced. Limits are set high enough to encourage dependency. During that phase, the strategic question is not how much value can be extracted from each user. It is how many users can be brought inside the system.

That phase appears to be ending.

OpenAI said in February 2026 that ChatGPT had reached 900 million weekly active users and 50 million paying subscribers. At that scale, the company is no longer merely proving demand. It is managing a market large enough for segmentation itself to become a core strategic lever.

Once a user base reaches that scale, the logic changes.

The question becomes how access is divided.

The Pricing Shift

The pricing shift is visible in two different forms.

Anthropic’s approach is the more direct one. Usage is increasingly tied to token consumption. Revenue rises with actual compute demand. Heavy use is not hidden behind a flat monthly abstraction. It is exposed and billed. OpenAI’s approach is more layered. It combines broad subscriptions with progressively higher access tiers and, in some products, movement toward token-based accounting as well. OpenAI’s help documentation says Codex pricing was updated in April 2026 to align with API token usage rather than per-message pricing for Plus, Pro, Business, and new Enterprise customers.

These are different tactical responses.

They point toward the same structural condition.

Agentic usage is expensive. Long-running workflows, parallel tasks, and deep reasoning loops consume far more compute than casual chat. A pricing model built for light conversational use becomes difficult to sustain once users begin treating frontier models as working systems rather than novelty interfaces. Anthropic’s public pricing for Claude Opus 4.6 and OpenAI’s tiered ChatGPT plans both make that pressure visible.

The token moat begins there.

It begins when frontier access is no longer priced mainly to attract users, but to sort them.

The Growing Divide

One consequence is easy to see.

The divide grows.

As stronger models, higher limits, longer contexts, deeper research tools, and more sustained agent execution move upward into higher-priced plans, the distance between those who can afford frontier usage and those who cannot begins to widen. That widening does not need to be mathematically exponential to be structurally significant. It is enough that it compounds.

A user with access to the best model does not merely get better answers.

That user gets more attempts. More failed experiments survived. More iterations completed. More search space explored. More parallel workflows sustained. More mistakes caught before they become costly.

In an agentic environment, these differences stack.

The result is that pricing no longer governs only quantity of use. It governs developmental velocity.

This is what makes the token moat more serious than an ordinary premium upgrade path. It is not just a matter of convenience. It is a matter of how much productive cognition can be purchased and sustained over time. OpenAI’s public plan structure already reflects large jumps in access between Free, Plus, and Pro, while Anthropic’s API pricing makes the cost of sustained advanced usage explicit.

The Small-Team Advantage

But that is not the whole picture.

The same structure that widens the divide also increases the leverage of small actors.

A single developer, or a pair of developers, with strong judgment and mastery of multi-agent workflows can now command an amount of software labor that would previously have required a much larger team. In that setting, the bottleneck shifts. Raw implementation ceases to be the main constraint. The main constraints become architectural clarity, orchestration skill, and budget discipline.

This is not a trivial shift.

It is liberating.

A $200 monthly budget for a frontier plan, or a few hundred dollars in carefully managed token spend, can now give a small operator access to an amount of synthetic labor that would once have been inaccessible. The human governance layer becomes more important precisely because the execution layer has become more fluid. OpenAI’s Pro tier is explicitly positioned around heavier use and substantially higher limits than Plus, while Anthropic’s pricing makes clear that the more capable agentic workflows remain available to those willing and able to pay for them.

That means the token moat has a dual effect.

It stratifies access. It also compresses the amount of scale required to matter.

Those outcomes are not contradictory.

They are two sides of the same transition.

The App Store Analogy

This pattern is not entirely new.

The early App Store also appeared radically flattening. Small developers could reach mass markets, and in some cases could compete with established firms on nearly equal footing. Distribution had been compressed. A single successful app could generate returns previously reserved for much larger organizations.

But that flattening was never unconditional.

It sat behind a gate: Apple hardware, Apple tooling, Apple rules, Apple economics. Over time, larger companies learned how to operate inside that gate more systematically. Individual success remained possible, but the broader relative advantage of scale reasserted itself. More resources still meant more development capacity, more marketing, more iteration, more resilience, and a better chance of success.

Frontier AI appears likely to follow a similar trajectory.

The initial effect is empowering because the technology genuinely increases the output of small teams. But once access to the strongest models becomes more metered and tiered, capital advantage begins to re-enter the system. The leverage of the individual rises in absolute terms, while the relative distance between large and small actors remains capable of widening.

That is not a failure of the technology.

It is a consequence of its pricing structure.

The Unknown Variable

This entire dynamic depends on one important uncertainty.

It depends on whether local and open systems remain materially behind the closed frontier.

If they do, the token moat hardens. The strongest closed systems remain the best way to purchase high-end machine cognition, and those who can afford more of that cognition continue to move faster.

If they do not, the moat weakens.

Google’s Gemma 4 release matters in this context because it shows that strong open models continue to improve, and Google said in April 2026 that Gemma downloads had passed 400 million. That is evidence of real diffusion. It is not yet proof of frontier equivalence. Gemma 4 may be highly capable while still remaining below the practical ceiling of the strongest closed offerings.

That distinction matters.

The token moat is not inevitable in the strongest sense. It is conditional on the persistence of closed-model superiority at the top end. If open or local stacks cross the threshold of practical equivalence for the most valuable workflows, then access-based stratification becomes harder to sustain. If they do not, then the moat remains economically meaningful.

The Structural Tension

This is why the token moat should not be understood as a simple story of exclusion.

It is more complicated than that.

Frontier AI may allow a single skilled developer to compete with a thirty-person software house in ways that were previously impossible. At the same time, the increasing use of tiered subscriptions, metered token billing, and higher-cost access paths may ensure that those with deeper pockets still ride the crest of the wave for longer.

The technology can flatten execution while preserving hierarchy.

It can reduce the scale required to compete while still increasing the value of being able to pay for more compute, more limits, and more continuity of access.

That is the real tension.

The token moat does not eliminate empowerment.

It conditions it.

Conclusion

The common intuition is that more capable models should make intelligence more abundant.

That intuition is incomplete.

What matters is not only how much intelligence exists, but how access to that intelligence is structured. Once frontier models move from reach-building to extraction, tokens cease to function merely as an internal billing unit. They become the mechanism through which access is divided, throttled, and priced.

That is the token moat.

Its effect is not singular.

It widens the divide between users who can afford sustained frontier access and those who cannot. It also gives small, highly skilled operators a level of leverage that would once have been unattainable. It may empower individuals in absolute terms while preserving or even amplifying structural advantage in relative terms.

Whether that moat endures depends on a separate question: whether local and open systems can narrow the gap to the point of practical irrelevance.

If they can, access becomes harder to monopolize.

If they cannot, then frontier AI will not simply be a story of intelligence becoming cheaper.

It will also be a story of intelligence becoming tiered.