With just days left in the Biden administration, the Commerce Department has published its most ambitious AI-related regulation yet: the AI Diffusion Rule, a sweeping framework that creates a three-tier system governing the export of advanced AI chips and model weights. Whether you’re building AI systems, deploying them internationally, or simply relying on cloud infrastructure, this rule will ripple through the industry for years.
The rule, which was published on January 15 and is set to take effect in 120 days, moves beyond the existing entity-specific export controls on China and creates a global framework. It’s arguably the most significant piece of technology policy to come out of this administration, and it arrives at a moment when the incoming Trump administration’s approach to AI regulation remains uncertain.
The Three-Tier Framework#
The rule divides the world into three tiers for purposes of advanced AI chip exports:
Tier 1 consists of 18 close allies and partners — including the EU, UK, Japan, South Korea, Australia, and others. These countries face essentially no restrictions. Companies headquartered in Tier 1 nations can purchase and deploy advanced GPUs (think NVIDIA H100s, A100s, and future generations) without meaningful limitations.
Tier 2 covers most of the rest of the world — roughly 120 countries. Companies in these countries can import a limited number of advanced AI chips (up to about 50,000 GPUs per company) without a license. Beyond that threshold, they need to establish “Universal Verified End Use” (UVEU) agreements with the US government, essentially promising the chips won’t be diverted.
Tier 3 is the restricted group: China, Russia, Iran, North Korea, and a handful of others. These countries remain under the tightest restrictions, with effectively no access to the most advanced AI chips.
The framework also introduces controls on closed-weight AI model exports. Models exceeding certain compute thresholds during training would require licenses for deployment in Tier 2 and Tier 3 countries. Open-weight models are explicitly exempted, which is a notable win for the open-source AI community.
Why This Matters for Developers#
If you’re a developer or architect working on AI deployments, the practical implications depend entirely on where you and your users are located.
For those of us in Europe, the impact is minimal — the Netherlands, along with the rest of the EU, sits comfortably in Tier 1. We can continue procuring and deploying whatever compute we need. But if you’re building services for international customers, particularly in Southeast Asia, the Middle East, Africa, or Latin America, you now need to think about the compute geography of your deployments.
The 50,000 GPU cap for Tier 2 countries is meaningful. That sounds like a lot, but for major cloud providers trying to build data center capacity in places like India, Brazil, or the UAE, it’s actually quite limiting. A single large training cluster can use 10,000 to 30,000 GPUs. This rule effectively constrains the build-out of AI compute capacity in Tier 2 nations unless those countries negotiate UVEU frameworks with the US.
The cloud implications are interesting too. The rule includes provisions for “headquarter-based” exemptions — if a Tier 1 company (say, Microsoft or Google) operates data centers in Tier 2 countries, they can receive higher allocations, but with conditions around security and access controls. This could give the US hyperscalers a structural advantage over local cloud providers in Tier 2 markets.
The Geopolitical Chess Game#
Let’s be honest about what’s really going on. This rule is primarily about China. The original chip export controls in October 2022 were targeted but leaky — Chinese companies found workarounds through third countries and slightly modified chip designs. The AI Diffusion Rule is an attempt to plug those gaps by controlling the entire global distribution, not just the China-facing exports.
The strategy is clear: rather than playing whack-a-mole with specific entities and chip models, create a comprehensive framework that controls the global flow of AI compute. It’s a recognition that in a world where AI chips are the new strategic resource, you need something more like an oil export regime than traditional tech export controls.
NVIDIA has already pushed back, arguing the rule could drive customers to foreign competitors and fragment the global tech ecosystem. Their concern isn’t unfounded — Huawei’s Ascend AI chips, while less capable than NVIDIA’s best, are being deployed at scale within China. If Tier 2 countries find US-made chips too difficult to procure, they might look to alternatives.
The Open-Source Exemption#
One detail that’s getting less attention than it deserves: the rule explicitly exempts open-weight model exports from the licensing requirements. This means models like Meta’s Llama, Mistral’s offerings, and other openly released models can be deployed anywhere in the world without additional export controls.
This is significant policy. It suggests the administration recognizes a distinction between compute infrastructure (which is physically constrained and controllable at borders) and model weights (which are essentially information and practically impossible to control once released). By exempting open-source models, the rule avoids the absurdity of trying to control the distribution of files that are already freely available on the internet.
For the open-source AI community, this is a meaningful victory. It preserves the ability to share AI research and models globally, even as the underlying compute infrastructure becomes more controlled.
My Take#
I have mixed feelings about this rule. On one hand, I understand the national security logic. Advanced AI capabilities are genuinely dual-use, and there are legitimate reasons to want to control which governments have access to the most powerful AI systems. The three-tier approach is more nuanced than a blanket ban.
On the other hand, as someone who’s spent decades in an industry built on global collaboration and open standards, I’m uncomfortable with the trajectory. The internet was designed to be borderless. Open source thrived because code could flow freely. Introducing export controls on compute and model weights — even well-intentioned ones — starts to balkanize the global technology ecosystem.
The practical question is whether the incoming Trump administration will keep, modify, or scrap this rule. The 120-day implementation timeline means it won’t take full effect until mid-May, giving the new administration ample time to intervene. Given the political dynamics around both China hawkishness and tech industry lobbying, the outcome is genuinely hard to predict.
What’s clear is that AI governance is no longer a theoretical discussion. It’s becoming trade policy, export control, and geopolitics. For those of us who just want to build useful software, the complexity of the operating environment just increased significantly.
