Computerworld NZ
Microsoft and OpenAI on Monday again revised their agreement, softening their exclusivity and revenue-sharing conditions in the process. These changes underscore how critical it is for enterprises to work with as many AI vendors as practical, given the leapfrogging performance stats as well as the constantly shifting alliances. Both OpenAI and Microsoft issued their own statements, which were essentially identical, about the contractual changes. Microsoft’s statement said that the company still derives some benefits from its alliance with OpenAI. “Microsoft remains OpenAI’s primary cloud partner and OpenAI products will ship first on Azure, unless Microsoft cannot and chooses not to support the necessary capabilities,” it said. But, the company noted, the earlier exclusivity is now gone. “OpenAI can now serve all its products to customers across any cloud provider. Microsoft will continue to have a license to OpenAI IP for models and products through 2032. Microsoft’s license will now be non-exclusive.” In addition, the company’s role as a major investor in OpenAI is driving a different revenue relationship, it said: “Microsoft will no longer pay a revenue share to OpenAI. Revenue share payments from OpenAI to Microsoft continue through 2030, independent of OpenAI’s technology progress, at the same percentage but subject to a total cap. ” AGI clause removed One key component within earlier versions of the Microsoft-OpenAI deal was the change in the relationship if OpenAI ever achieved artificial general intelligence (AGI), a term that eludes a concrete definition but generally refers to AI that equals or exceeds human capabilities. Although it was not directly referenced in the statement from either vendor, multiple media reports said that AGI references have now been removed from the revised agreement. Market changes Analysts and consultants generally agreed that this altered agreement will reinforce, and should extend, the current enterprise IT trend of hedging bets by striking arrangements with a variety of AI providers, including the major hyperscalers. Beyond future-proofing enterprises’ AI efforts, some of those agreements are for practical issues, such as the need to work with global AI firms specializing in different languages that the enterprise needs . Thomas Randall , research director at Info-Tech Research Group, explained that the market has changed since the original agreement was struck. “The era of exclusive frontier model access as a strategic differentiator is coming to an end,” he pointed out. “The Microsoft-OpenAI agreement in 2023 was meaningful because access to GPT4 was scarce. But that scarcity no longer applies because the competitive differences between frontier models have reduced substantially since then.” The amended Microsoft-OpenAI agreement “is more of a formal acknowledgment that model access is no longer a strict advantage,” he said. “The immediate practical change for IT from this agreement, especially for shops that were reluctant to deepen an Azure commitment, is that they now have a clearer path to accessing OpenAI models through other hyperscalers.” Randall argued that this translates into a rebalancing of where enterprise IT should focus its AI efforts, especially in terms of differentiation. “If model access is commoditizing at the infrastructure layer, then strategic questions must focus on quality and governance of proprietary data, the depth and sophistication of agentic workflow integration, and organizational capability to deploy AI at scale,” he said. “Consequently, the vendors who control the orchestration and application layers [such as] the agent frameworks, the data connectors, the governance tooling, and workflow integration, will be best positioned to capture enterprise value. The competitive ground has shifted from attaining model access to how vendors deeply and reliably embed AI into enterprise workflows.” Alastair Woolcock , VP analyst at Gartner, agreed that this contractual change from two key market leaders is an inevitable reaction to a vastly changing AI marketplace. “The first great AI shadow investment is being rewritten for a multipolar AI Cold War,” he said. “Frontier AI has become too capital-intensive and infrastructure-constrained for one-cloud exclusivity to survive. For Microsoft, this is a controlled concession. The investor story moves from ‘Microsoft owns the OpenAI channel’ to ‘Microsoft controls the enterprise AI operating layer’ through Copilot, Azure, security, workflow integration, data gravity and AI operations,” Woolcock said. “For OpenAI, this is a liberation event,” he noted. “Its biggest constraint is no longer demand. It is compute, capital and distribution. OpenAI cannot become the global AI platform if one partner controls the pipes.” He added that, for enterprise IT executives, “this means more choice, but not necessarily less dependency. Lock-in moves up the stack, from cloud infrastructure to AI ecosystem alignment, agent orchestration, workflow control and data governance. This is consequential, not because the partnership is weakening, but because it shows the next phase of AI competition will be fought through flexible alliances, compute access, silicon, power and enterprise distribution, not traditional ownership.” Planning assumptions altered Tony Olvet , group VP with IDC, said this contractual change “is unlikely to affect most near‑term Microsoft or OpenAI deployments, but it does change planning assumptions. CIOs and CTOs should expect more choice in where OpenAI capabilities appear, greater commercial leverage and increased need to govern AI across multiple channels. This has strategic implications: enterprises should continue to rely on strong partners while designing AI architectures, contracts, and governance frameworks that can shift across clouds, models and vendors as the market evolves.” Most consultants stressed the vanishing exclusivity for almost all of the key AI players, something that may not be a bad thing for IT. A key background factor at play here is the timeline. It can take an enterprise an extended period to fully deploy capabilities across its global environment. Noah Kenney , principal consultant for Digital 520, noted, “standing up OpenAI workloads on AWS, Google Cloud, or Oracle will take time. Reference architectures, identity and data integrations, compliance reviews, and procurement cycles do not move at the speed of a press release. Enterprises that have spent years optimizing on Azure will not migrate overnight, nor should they.” But, he said. “for the substantial population of companies that are not Microsoft shops, that have actively avoided Azure, or that operate in multi-cloud by policy, this is the first time OpenAI has been a realistic first-class option on their preferred infrastructure. That is a meaningful shift in the addressable market, even if the operational reality lags by quarters.” Given the constantly changing relationships within AI, not to mention multiple AI firms preparing to try to become publicly traded, reality is likely to look very different at the end of an enterprise AI rollout than it did at the beginning, so they need options. “Until today, choosing OpenAI effectively meant choosing Azure, and choosing Azure gave you privileged access to OpenAI. That tight coupling shaped procurement decisions, reference architectures, and multi-year cloud commitments at thousands of enterprises. It is no longer true,” Kenny said. “What changes for [enterprise IT executives] is the structural assumption underneath their AI roadmap,” he noted. “OpenAI can now ship its products across any cloud and Microsoft now has a non-exclusive license to OpenAI’s IP through 2032, which means Microsoft is also free to lean harder into its own models, into Anthropic, and into whatever else the market produces. Both sides just bought themselves optionality and that optionality flows downstream to the customer.” He added, “the companies that benefit are the ones who treat model providers, cloud providers, and inference infrastructure as three separate procurement decisions with three separate exit ramps.” Vendor lock-in ‘relocating’ Sanchit Vir Gogia , chief analyst at Greyhound Research, said that the kneejerk reaction to the contract changes is that enterprise IT will now have more options and more flexibility. But Gogia said that dependence is not being reduced as much as it is being moved. “Lock-in is not going away. It is relocating. At the model level, substitution is becoming easier. Not trivial, but certainly more feasible than before. At the orchestration level, however, substitution remains difficult,” Gogia said. “Once your workflows, controls, identity layers, and governance structures are built around a particular system, changing that system is not a small task. That is where dependency sits. Quietly. Persistently. And often unnoticed until it begins to constrain you.” There are still differences between providers, and those differences matter in certain contexts, he said. “But the gap is narrowing in ways that are meaningful for enterprise use. Increasingly, the question is not which model is best in isolation. The question is how that model is used, governed, and embedded into the organization. That is a very different question,” Gogia said. And, he pointed out, it leads you to a very different place, “because once you ask that question, you are no longer looking at models. You are looking at orchestration. You are looking at identity. You are looking at governance, compliance, integration, workflow. You are looking at the layer that sits above the model and quietly determines how everything actually works. That layer is where the real dependency forms.” Microsoft understands this, he noted. “You can see it in how it is positioning itself. It is no longer behaving like a gateway to a single provider. It is building something broader: A layer where multiple models can coexist, where those models can be managed, governed, and embedded into enterprise systems in a consistent way. That is not accidental,” Gogia said. “That is a deliberate move towards control at a higher level. And importantly, it is also a hedge. A very clear one. Because it reduces reliance on any single partner, including OpenAI.”
Go to News Site