The Cloud AI Marketplace War Is Over, and AWS Won
Amazon Web Services effectively won the cloud AI marketplace war by becoming the neutral aggregation layer where every major model provider, including the formerly Microsoft-exclusive OpenAI, now competes for enterprise customers. AWS's strategy of hosting rivals rather than betting on a single model transformed Bedrock into the default procurement destination for enterprise AI workloads.
By Andrew Baker | andrewbaker.ninja
On 28 April 2026, something quietly significant happened to the enterprise technology landscape. OpenAI’s models went live on Amazon Bedrock, less than 24 hours after the company’s seven-year Microsoft Azure exclusivity arrangement formally expired. The move was announced at an AWS event in San Francisco, and it was framed, accurately, as meeting years of pent-up customer demand. But the headline undersells what actually happened. This was not simply a distribution deal. It was the moment Amazon Web Services completed its transformation from cloud infrastructure provider into the default operating layer for enterprise artificial intelligence.
1. How We Got Here
The Microsoft and OpenAI relationship was, for a long time, the defining partnership in enterprise AI. It began in 2019 when Microsoft committed $1 billion to OpenAI in exchange for the right to host its models on Azure. That arrangement was reinforced in 2023 with a $10 billion investment and deepened further through tight integrations with Copilot, GitHub, Bing, and Microsoft 365. For enterprise customers, the message was unambiguous: if you wanted access to the best frontier AI, you went through Azure.
AWS spent years watching that dynamic with quiet frustration. Its answer was not to build a competing frontier lab. It was to build a better shelf. Amazon Bedrock launched as a managed inference platform designed around the principle that customers should be able to choose the best model for every use case, accessed through a single, consistent API with unified security, governance, and cost controls. That is a deceptively powerful idea. Rather than betting on a single model winning, AWS bet on the catalogue being the competitive advantage.
The strategy required patience and capital. AWS invested up to $25 billion in Anthropic, locking in Claude as a cornerstone tenant. It brought in Meta, Mistral, Cohere, and its own Nova family. It built AgentCore, PrivateLink integration, IAM-based access management, and CloudTrail logging into the platform. By the time OpenAI’s exclusivity with Microsoft expired on 27 April 2026, Bedrock was already serving more than two million developers across more than a hundred foundation models. All AWS needed was the final piece.
2. What the OpenAI Deal Actually Means
The announcement covered three distinct capabilities, all currently in limited preview. OpenAI’s frontier models, including GPT-5.5 and GPT-5.4, are now available on Bedrock through the same APIs customers already use for every other model. Codex, OpenAI’s coding agent, can be configured to route inference through Bedrock infrastructure, allowing enterprise teams to authenticate using their existing AWS credentials and apply usage toward their existing cloud commitments. And Amazon Bedrock Managed Agents, powered by OpenAI, gives enterprises a platform for building production-ready agentic applications with OpenAI models running inside AWS’s security and compliance envelope.
That last point is the one most likely to get glossed over. The managed agents platform handles orchestration, tool use, memory across calls, and governance, the hard parts of moving from a prototype to something that runs reliably in a production environment. OpenAI’s chief revenue officer Denise Dresser said publicly that inbound demand for the AWS offering had been, in her word, staggering. AWS CEO Matt Garman was more direct: “Their production applications run in AWS. Their data is in AWS. They trust the security of AWS, and we’ve forced them for the last couple of years, to get great OpenAI models, to go to other places.”
That quote is worth sitting with. AWS was not describing a new market it had created. It was describing a constraint it had finally removed. The customers were already there. The workloads were already there. The data was already there. The AI just needed to follow.
3. What This Means for Microsoft
It would be easy to read this as a catastrophic loss for Microsoft, but the reality is more nuanced. Azure remains OpenAI’s primary cloud provider under a licensing arrangement that runs through 2032. Azure OpenAI Service retains deep integrations with Microsoft 365 Copilot, GitHub Copilot, Dynamics 365, and Microsoft Security Copilot, integrations that Bedrock cannot replicate without substantial parallel investment. And Microsoft has been quietly repositioning for exactly this outcome, with its MAI in-house model programme and an expanded Anthropic partnership that now puts Claude across multiple Copilot surfaces alongside its own models.
Satya Nadella told analysts on Microsoft’s January 2026 earnings call that diversification of model providers, including its own, was a strategic priority. Microsoft has been building the exits before the house caught fire. The loss of exclusivity was not a surprise event. It was the culmination of a renegotiation in which Microsoft traded its lock-in rights for relief from revenue share commitments. Both companies got something they needed. Microsoft got a cleaner path to building its own model stack. OpenAI got the freedom to go wherever enterprise customers actually live.
4. The Azure Foundation Problem Nobody Wanted to Talk About
The official narrative frames OpenAI’s move to AWS as a commercial decision about enterprise distribution. A more honest reading suggests it was also a vote of no confidence in Azure’s technical foundations, and that the warning signs have been visible for years to anyone paying attention.
Axel Rietschin, who worked as an Azure Core Compute engineer and spent eight years before that on the Windows Base Kernel team, has been publishing a detailed series of essays on exactly this subject. His account, written from direct experience, describes a platform that was rushed to market in 2008 to compete with AWS and never fully recovered from the decisions made in that sprint. He argues that Microsoft’s rushed Azure launch, the talent exodus that followed, and persistently poor execution left the service in a state of permanent triage. As he put it, Azure never operated as smoothly or independently as promised, and what Microsoft presented to its most demanding customers was a sophisticated system on permanent life support.
The detail that most struck me in Rietschin’s account was not the architectural chaos, though that is striking enough. It was the 173 agents. On his first week at Azure Core in 2023, he discovered that the team responsible for managing each Azure node had identified 173 software agents as candidates for porting to new hardware, and that nobody in the organisation could explain what all of them did, how they interacted, or why they existed. That is not a software quality problem. That is an organisational knowledge collapse, the kind that accumulates silently over years of attrition, rushed hiring, and management layers that reward shipping over understanding.
Rietschin ties this directly to OpenAI’s $11.9 billion compute deal with CoreWeave in March 2025, which he reads as a signal that Microsoft was struggling to meet OpenAI’s infrastructure requirements at scale and on time. The Register’s coverage of the Rietschin essays connects this further to Microsoft’s layoff of around 15,000 people during the May to July 2025 period, and to GitHub’s well-documented availability problems as it migrated traffic onto Azure infrastructure. The pattern is consistent: a platform under strain, staffed by an organisation that has been systematically thinned, attempting to absorb one of the highest-throughput AI workloads in the world.
None of this means Azure is broken in ways that matter for routine enterprise workloads. Most organisations running general-purpose compute on Azure will see none of this. But for the specific challenge of serving frontier AI inference at the scale OpenAI requires, with the reliability and governance that enterprise customers demand, the cracks were apparently real enough to drive a $38 billion commitment to a competitor. That is the part of the story the press release version of this week’s announcement leaves out.
5. The Infrastructure Thesis Is Winning
There is a broader argument here that extends well beyond the OpenAI deal, and it is one that I think gets insufficient attention in most coverage. The real story of AI in 2026 is not which model is best. It is that AI winners are increasingly being decided by infrastructure depth, not model quality alone.
Amazon is on pace to spend $200 billion in capital expenditure this year, the vast majority of it on AI infrastructure. Its custom silicon business, Trainium and Inferentia, is generating more than $20 billion a year in revenue. The $38 billion, seven-year compute commitment that OpenAI signed with AWS in November 2025 requires OpenAI to spin up two gigawatts of Trainium accelerators. That is not a software deal. That is a physical infrastructure bet that ties the most important AI company in the world to Amazon’s datacentre footprint for nearly a decade.
Google, for its part, launched its eighth-generation Tensor Processing Units in late April 2026, splitting the architecture into two specialised chips: the TPU 8t for large-scale training, and the TPU 8i optimised for high-volume inference workloads. The inference chip in particular is designed to lower the marginal cost of running AI at scale, which matters enormously as the industry shifts from experimentation to production. Both AWS and Google are making the same underlying bet: that vertical integration in silicon is the most durable competitive advantage in the AI era, and that whoever controls the compute controls the margin.
Microsoft, notably, is further behind in custom silicon than either of its major rivals. That asymmetry may matter more over the next five years than any individual model release.
6. What This Means for Enterprise AI Buyers and African Banks Specifically
For most enterprise technology leaders, the practical implication of this week’s news is straightforward: your AI procurement decision just got simpler, and your negotiating position just got stronger.
If you are already an AWS customer, and the vast majority of large enterprises in Africa are, you can now access GPT-5.5, Claude, Llama, Mistral, and Cohere through the same IAM policies, the same PrivateLink configuration, the same compliance controls, and the same cloud commitment you already have in place. There is no new security model to learn, no new vendor contract to negotiate, no new data residency conversation to have with your information security team. That friction reduction is significant. In large financial institutions, the gap between “we could technically use this model” and “we have approved this model for production use” is often measured in months. Bedrock’s unified governance layer compresses that gap considerably.
For South African banks in particular, this shift has a specific flavour. The conversation about AI adoption in local financial services has long been constrained by two things: data sovereignty concerns and the difficulty of integrating frontier models with legacy risk and compliance infrastructure. Bedrock’s architecture directly addresses both. Your data stays inside the AWS region, af-south-1 in most cases, processed under your existing contractual and regulatory framework. Your existing CloudTrail audit logs, your existing IAM governance, your existing PrivateLink connectivity: all of it extends to cover OpenAI and Anthropic models simultaneously.
The question for South African CIOs is no longer which cloud to use for AI. It is which models to prioritise, which workloads to automate first, and how to build the internal muscle to evaluate and govern AI outputs at scale. Those are genuinely hard questions, but they are better questions than the ones the industry was asking eighteen months ago.
7. The Bedrock Catalogue as Strategic Moat
What Amazon has built with Bedrock is, in infrastructure terms, something analogous to what it built with the original AWS service catalogue in the mid-2000s. The genius of early AWS was not any individual service. It was the combination of commoditised building blocks, pay-as-you-go pricing, and a unified developer experience that made the whole more valuable than the sum of its parts. Each new service that joined the catalogue made the platform stickier, reduced the incentive to go elsewhere, and generated data that helped Amazon improve the platform further.
Bedrock is following the same pattern one level up the stack. Each new model that joins the catalogue increases the probability that any given enterprise workload can be served from within the platform. Each new enterprise control, guardrails, watermarking, latency routing, model evaluation, makes it harder to justify running AI workloads outside of it. And each dollar of cloud commitment that gets applied to AI model usage is a dollar that does not go to a standalone API contract with a model provider, which means AWS is extracting value from the AI layer without having to win the model race itself.
This is, I think, the deepest reason why this week matters. OpenAI joining Bedrock is not primarily a story about OpenAI gaining distribution. It is a story about AWS demonstrating that its catalogue strategy works, that the gravity of an established enterprise cloud platform is strong enough to pull even the most consequential AI companies into its orbit. If GPT-5.5, Anthropic’s Claude, and Meta’s Llama all sit on the same shelf behind the same API, behind the same IAM policy, billed to the same cloud commitment, then the cloud platform is the product. The models are features.
8. Where This Leaves Everyone
AWS is, as of this week, the default enterprise AI platform. It did not get there by building the best model. It got there by building the best infrastructure for running every model, and by being patient enough to wait for the model wars to resolve into a multi-provider world.
Microsoft retains meaningful advantages in the productivity and developer tooling stack, and its long-term Azure relationship with OpenAI preserves significant enterprise entrenchment. Google Cloud has the best custom silicon story outside of AWS, and its vertical integration from TPU to model to application layer is a credible counterweight. But in terms of raw enterprise AI platform gravity, measured by the breadth of models available, the depth of governance controls, and the number of existing enterprise workloads already running on the platform, AWS has a lead that will be difficult to close.
For enterprise AI practitioners, the message is clear. Stop optimising for model selection. Start optimising for your platform posture. The teams that will extract the most value from AI over the next three years are the ones that build durable internal capabilities: evaluation frameworks, governance tooling, agent orchestration patterns, and the organisational muscle to move from prototype to production quickly. The models will get better on their own. The infrastructure is here. The question now is purely execution.
References
- AWS and OpenAI, “AWS and OpenAI announce expanded partnership to bring frontier intelligence to the infrastructure you already trust,” About Amazon, 28 April 2026. https://www.aboutamazon.com/news/aws/bedrock-openai-models
- OpenAI, “OpenAI models, Codex, and Managed Agents come to AWS,” OpenAI Blog, 28 April 2026. https://openai.com/index/openai-on-aws/
- Todd Bishop, “OpenAI’s models land on Amazon Bedrock, one day after Microsoft exclusivity ends,” GeekWire, 28 April 2026. https://www.geekwire.com/2026/openais-models-land-on-amazon-bedrock-one-day-after-microsoft-exclusivity-ends/
- Jordan Novet, “OpenAI brings models to AWS after ending exclusivity with Microsoft,” CNBC, 28 April 2026. https://www.cnbc.com/2026/04/28/openai-brings-models-to-aws-after-ending-exclusivity-with-microsoft.html
- “OpenAI jumps out of Microsoft’s bed, into Amazon’s Bedrock,” The Register, 28 April 2026. https://www.theregister.com/2026/04/28/openai_climbs_into_amazons_bedrock/
- “OpenAI on AWS Bedrock: $38B Deal Ends Azure Lock-In,” Tech Insider, 28 April 2026. https://tech-insider.org/openai-amazon-bedrock-38-billion-azure-exclusivity-end-2026/
- AWS, “Top announcements of the What’s Next with AWS, 2026,” AWS Blog, 29 April 2026. https://aws.amazon.com/blogs/aws/top-announcements-of-the-whats-next-with-aws-2026/
- “Top Tech News Today, April 28, 2026,” Tech Startups, 28 April 2026. https://techstartups.com/2026/04/28/top-tech-news-today-april-28-2026/
- Thomas Claburn, “Ex-Microsoft engineer believes Azure problems stem from talent exodus,” The Register, 4 April 2026. https://www.theregister.com/2026/04/04/azure_talent_exodus/
- Axel Rietschin, “How Microsoft Vaporized a Trillion Dollars,” Axel’s Substack, 29 March 2026. https://isolveproblems.substack.com/p/how-microsoft-vaporized-a-trillion