Moon Foundry Moon Foundry
All essays
AI & Marketing

The Infrastructure You Cannot See Is the Infrastructure That Owns You

The most consequential decisions in AI are not being made by the people building on top of it.

Alex Albano | | 8 min read

In 1995, if you wanted to build a website, you needed to understand HTTP, DNS, TCP/IP, and how to configure a web server. The infrastructure was visible because you had to touch it. By 2005, shared hosting and early cloud platforms had abstracted most of that away, and you could build a web application without understanding the protocols it ran on. By 2015, you could deploy a global application on AWS without knowing which data centres it operated in, which routing decisions determined how traffic reached your users, or which upstream capacity constraints shaped your application’s performance characteristics. The infrastructure had become invisible, and the invisibility was the product. The people building on top of the stack could move faster precisely because they did not have to think about what sat beneath them.

The same pattern is unfolding in AI, faster than the internet precedent and with higher structural stakes, because the decisions being made at the infrastructure layer encode constraints that are more consequential and less reversible than those of the cloud era. And for marketing practitioners specifically, the pattern is worth understanding because the marketing stack has become one of the most AI-dependent operational layers in most organisations, which means the infrastructure decisions being made underneath it are shaping what marketing can do, what it costs, and what it is allowed to say.

The stack as argument

When a marketing team builds an agentic workflow using an API from OpenAI, Anthropic, Google, or any of the current foundation model providers, they are making a series of decisions that feel like tool selection and that function as structural commitments.

The model they choose determines the ceiling on what their system can reason about, the floor on what it costs to operate, and the boundaries of what their system can say. Every foundation model carries embedded assumptions about safety, alignment, and acceptable use that shape the output in ways the downstream builder does not control and usually does not fully understand. A content generation pipeline built on one provider’s model will produce systematically different output from the same pipeline built on another, the same brand brief and the same audience data yielding different tonal ranges, different degrees of creative risk, different comfort with competitive claims, because the infrastructure layer carries its own arguments about what kind of content is acceptable and where the boundaries of autonomous expression should sit. Any marketer who has switched between models mid-project and noticed the output shift in ways they could not fully explain has encountered this dynamic without necessarily naming it.

The pricing structure they accept determines the economics of their marketing operations at a level they cannot easily change. API pricing is not just a cost input. It is a design constraint. A personalisation workflow that makes fifteen agent calls per prospect interaction, enrichment, scoring, content generation, tone adjustment, compliance check, is economically viable at one price point and structurally unaffordable at another, which means that pricing decisions made by foundation model providers shape the architecture of every marketing workflow built on top of them. When a provider changes its pricing, every downstream team adjusts their architecture, whether they recognise the adjustment as an architectural change or experience it as a budget conversation in the quarterly review.

The context window they operate within determines what their system can hold in mind at any given moment, which determines what kinds of marketing problems it can address in a single pass. A campaign brief that includes brand guidelines, audience research, competitive positioning, tone of voice documentation, and performance data from previous campaigns may exceed the effective context window of the model processing it, producing output that draws on some of those inputs and silently drops others. Context window sizes are expanding, and they remain a binding constraint, and the marketing practitioners who understand this constraint design their workflows around it while the practitioners who do not encounter it as mysterious inconsistencies in output quality that they attribute to the tool rather than to the infrastructure underneath it.

Each of these infrastructure-layer decisions is an argument about what kind of AI products are possible, economically viable, and architecturally sound. The practitioners building on top of the stack are reading those arguments whether they know it or not. The ones who cannot read them are making structural commitments they do not understand.

The precedent is the internet, and the stakes are higher

The internet’s infrastructure battles produced lasting structural consequences that are still shaping the digital economy three decades later. The browser wars of the 1990s determined which rendering standards became dominant, which in turn determined what could be built on the web, which in turn shaped the economics of web-based businesses for twenty years. Google’s search monopoly, established in the early 2000s through a combination of technical superiority and network effects, created an advertising-funded infrastructure layer that shaped the business model of most of the internet’s content economy and that every marketer alive today has had to build their strategy around. AWS, by offering cloud infrastructure at scale, made a specific set of architectural choices, about how compute should be priced, how data should be stored, how applications should be deployed, that became the default assumptions of an entire generation of software developers.

In each case, the infrastructure-layer decisions were made by a small number of actors, adopted by a large number of builders who needed the infrastructure to function and did not have the resources or inclination to examine its embedded assumptions, and then locked in through switching costs that grew proportional to the ecosystem built on top of them. By the time the downstream builders understood the constraints they had accepted, the cost of changing those constraints was prohibitive.

The AI infrastructure layer is following the same pattern with two differences that raise the stakes considerably. First, the concentration is more extreme. The foundation model layer is dominated by fewer than ten organisations globally, and the effective choice for most practitioners is between three or four providers whose technical capabilities, pricing structures, and usage policies differ in ways that matter but that most downstream builders are not equipped to evaluate. Second, the decisions being made at this layer are not just about compute and storage. They are about reasoning, judgment, and the boundaries of autonomous action, which means the infrastructure is encoding assumptions about cognition itself, about what kinds of thinking are appropriate, what kinds of content are safe, and what kinds of autonomy are acceptable.

What the practitioner cannot see

The practical consequence for the practitioner building an AI-native product, workflow, or business is that the most important decisions affecting their system have already been made by someone else, and the mechanisms through which those decisions reach them are designed to be invisible.

When a foundation model refuses a specific kind of request, the practitioner encounters what feels like a capability limitation, a boundary that presents itself as technical fact while functioning as a policy decision made by the model provider about what their infrastructure should and should not do, encoded in the model’s training and reinforcement and transmitted to every downstream application as a hard constraint. The practitioner adapts their product to work within the constraint, usually without realising that they are adapting to a policy choice rather than a physical limitation.

When an API provider deprecates a model version, changes its pricing tier, or modifies its usage policies, the practitioner encounters what registers as a business environment change, a routine adjustment, while the underlying reality is a unilateral architectural decision by their infrastructure provider that reshapes the economics and capabilities of their product without their input or consent. The practitioner adjusts, absorbs the cost, and continues building, usually without accounting for the fact that the adjustment was forced by a dependency they chose but do not control.

When a context window expansion makes new kinds of applications possible, the practitioner encounters what presents as a capability improvement, a rising tide, while the mechanism underneath is an infrastructure provider’s decision about resource allocation and market positioning that happens to benefit certain downstream use cases, made for reasons that may have nothing to do with those use cases, and reversible at any point through pricing changes that make the expanded capability economically impractical for the applications that depend on it.

Reading the stack

The thesis that has been running through everything I write applies here with particular directness: every AI stack is an argument about who matters. The infrastructure layer’s argument is that the builders of foundation models matter most, because their decisions propagate downward through every product, workflow, and business built on top of them. The application layer’s argument is that the orchestrators matter most, because they translate raw capability into value. The user’s argument, usually implicit, is that the person with the problem matters most.

These arguments are not reconcilable, and they do not need to be. What matters is that the people building in this space can read them. A practitioner who understands that their API provider’s pricing is a design constraint, that their model’s safety boundaries are policy choices, that their context window is a resource allocation decision rather than a law of physics, is a practitioner who can build with structural awareness. They can diversify their infrastructure dependencies, architect for portability, design their products to survive the infrastructure changes that will inevitably come.

A practitioner who cannot read the stack is a practitioner who is building on someone else’s argument without knowing what it says, and who will discover the constraints only when the argument changes and their product breaks in ways they did not anticipate, because the breakage is coming from a layer they never thought to examine.

The infrastructure you can see is the infrastructure you can negotiate with. The infrastructure you cannot see is the infrastructure that sets your terms. For marketing practitioners building AI-native workflows, the most important skill right now might not be the ability to prompt, to orchestrate, or to design agentic campaigns. It might be the ability to look at the stack underneath the martech layer and read, clearly and without illusion, the argument it is making about what your marketing is allowed to become. The marketers who built their entire strategy around Google’s search infrastructure without understanding Google’s incentive structure learned this lesson the hard way, repeatedly, across a decade of algorithm changes. The AI infrastructure layer is encoding a similar set of constraints, with similar invisibility, and the marketing practitioners who learn to read those constraints now will be the ones who are not surprised when they shift.


Stay in the loop

Essays on growth strategy, brand, and building at the frontier. No spam.