<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Yathārth]]></title><description><![CDATA[AI Safety | Policy & Governance | Philosophy, Economics and History.]]></description><link>https://rudraaaspeaks.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!0q9d!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Frudraaaspeaks.substack.com%2Fimg%2Fsubstack.png</url><title>Yathārth</title><link>https://rudraaaspeaks.substack.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 08 May 2026 18:57:42 GMT</lastBuildDate><atom:link href="https://rudraaaspeaks.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Yatharth Maheshwari]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rudraaaspeaks@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rudraaaspeaks@substack.com]]></itunes:email><itunes:name><![CDATA[Yathārth]]></itunes:name></itunes:owner><itunes:author><![CDATA[Yathārth]]></itunes:author><googleplay:owner><![CDATA[rudraaaspeaks@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rudraaaspeaks@substack.com]]></googleplay:email><googleplay:author><![CDATA[Yathārth]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Collapse of ‘Strategic’ Voluntary AI Governance!?]]></title><description><![CDATA[The first real stress-test of corporate AI safety commitments didn't just fail; it revealed that the very assumptions undergirding international coordination are built on sand.]]></description><link>https://rudraaaspeaks.substack.com/p/the-collapse-of-strategic-voluntary</link><guid isPermaLink="false">https://rudraaaspeaks.substack.com/p/the-collapse-of-strategic-voluntary</guid><dc:creator><![CDATA[Yathārth]]></dc:creator><pubDate>Mon, 02 Mar 2026 13:53:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!btBs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>TL;DR</h1><p>For the past couple of months I&#8217;ve been very keen and excited about international coordination for AI Safety, have come across some thought provoking literature(s) like Forethought&#8217;s &#8220;<strong>The International AGI Project Series,</strong>&#8221; of which &#8220;<a href="https://www.forethought.org/research/intelsat-as-a-model-for-international-agi-governance">Intelsat as a Model for International AGI Governance</a>&#8221; became my initial read, following which, came across the <a href="https://arxiv.org/abs/2310.09217">MAGIC consortium paper</a>, <a href="https://arxiv.org/html/2507.06379v1">Belfield's four-institution framework</a>, <a href="https://internationalaisafetyreport.org/sites/default/files/2026-02/international-ai-safety-report-2026.pdf">International AI Safety Report 2026</a>. Following my AGI Strategy course, I was enrolled in the Frontier AI Governance Course, as the following to understand and dive more into Governance perspectives, as I did start focusing on certain aspects of International Coordination and Verifications right from the AGI strategy course (my previous work on <a href="https://rudraaaspeaks.substack.com/p/a-critical-assessment-of-the-iaea">A Critical Assessment of the &#8220;IAEA for AI&#8221; Model</a> which aimed to assess the analogy between IAEA and similar adaption for AI and the aspects that we might have taken into consideration for governing frontier AI.</p><p>Then, in the last week of <strong>February 2026</strong>, the US government designated Anthropic a &#8220;Supply-Chain Risk to National Security&#8221; for refusing to remove two safety guardrails from a Pentagon contract. And I realised that every major international coordination proposal I&#8217;d been studying rests on an assumption that just got demolished in real time. </p><p>The key assumptions that we&#8217;ve been relying on are that: <strong>the actors who need to coordinate&#8212;governments, frontier labs, international institutions&#8212;share a baseline commitment to the idea that AI safety constraints are legitimate</strong>.</p><p>In this piece I&#8217;ll be using the Anthropic&#8211;Pentagon confrontation as a case study to examine four questions: (a) what this does to domestic US AI governance; (b) how it reshapes the interaction between AI companies and governments on safety; (c) how other nations will likely respond to this case/takeaways; and (d) whether this accelerates or decelerates international coordination efforts. I have two competing hypotheses on that last question, and have my uncertainty which side would dominate.</p><p>ps: To explore the Anthropic vs DoW in depth, I&#8217;d highly recommend reading Zvi Mowshowitz&#8217;s <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Don't Worry About the Vase&quot;,&quot;id&quot;:573100,&quot;type&quot;:&quot;pub&quot;,&quot;url&quot;:&quot;https://open.substack.com/pub/thezvi&quot;,&quot;photo_url&quot;:null,&quot;uuid&quot;:&quot;8d18a8b1-1f4b-44dc-87e5-3a060673a0c1&quot;}" data-component-name="MentionToDOM"></span> articles (<a href="https://thezvi.substack.com/p/anthropic-and-the-department-of-war">part-1</a>, <a href="https://thezvi.substack.com/p/anthropic-and-the-dow-anthropic-responds">part-2</a>). </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!btBs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!btBs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!btBs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!btBs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!btBs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!btBs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2743543,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://rudraaaspeaks.substack.com/i/189524809?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!btBs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!btBs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!btBs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!btBs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3383b7d7-a6f3-4cd2-8529-71dc13b84381_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>In theory, global AI safety is a shared goal. In practice, it is crossed out first and reclassified as a liability. (via: ChatGPT)</em></figcaption></figure></div><h1>Background</h1><p>I wrote my first piece on the IAEA as a model for AI governance focusing specifically on what the safeguards regime&#8217;s structural failures indicate us about designing verification mechanisms for frontier AI systems.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> In that piece I broadly focused about institutions while in this one I intend to explore <em><strong>the political conditions</strong></em> that make institutions possible in the first place.</p><p>Exploring some of the existing literature like the Forethought Research series on the International AGI Project<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>,  introduces the idea of centralizing AGI development under a coalition of democratic countries to control the pace of progress and enable a global pause if necessary. Similarly, the <strong>MAGIC consortium proposal</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> outlines a "CERN-like" multilateral project designed to prevent a corporate race to the bottom by pooling sovereign resources into a single, safety-first research environment. Building on this, <strong>Haydn Belfield&#8217;s work</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a><strong> on compute governance</strong> envisions an "IAEA for AI" that uses the physical tracking of high-end chips as a verification tool to ensure no actor can secretly train models beyond agreed-upon safety limits. </p><p>Every proposal or a framework convention triggered by capability thresholds, or a bottom-up supervisory coordination scheme would require at minimum that the participating governments treat frontier AI safety as a <em><strong>legitimate policy objective</strong></em> rather than a strategic inconvenience to be coerced away, which is exactly what has happened in the case of Anthropic vs Pentagon, a sense of strategic inconvenience emerged. </p><div><hr></div><h1>The Timeline, Compressed</h1><p>Highlighting the chain of events in brief with facts:</p><blockquote><p>In July 2025, Anthropic signed a contract with the Department of Defense, reportedly worth around <strong>$200 million</strong> and the contract included two explicit safety guardrails: a) <strong>Anthropic's models would not be used for mass domestic surveillance</strong>, and b) <strong>they would not be used for fully autonomous weapons systems</strong>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> These weren't unusual restrictions rather they reflected the company's Responsible Scaling Policy and were consistent with Anthropic's public AI Safety Levels (ASL) framework.</p></blockquote><blockquote><p>On February 24, 2026, Defense Secretary Pete Hegseth met with Anthropic CEO Dario Amodei and demanded the removal of both restrictions and the Pentagon set a deadline: 5:01 PM, Friday, February 28.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p></blockquote><blockquote><p>On February 26, Amodei issued a public statement. The key sentence: "Domestic mass surveillance and fully autonomous weapons are uses that are simply outside the bounds of what today's technology can safely and reliably do", and added: "We cannot in good conscience accede to their request."<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p></blockquote><blockquote><p>On February 27, three things happened in rapid succession, </p><p>a) Trump issued an executive order directing all federal agencies to cease using Anthropic's products and services<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>,<br>b) Hegseth formally designated Anthropic a "Supply-Chain Risk to National Security," a legal classification that restricts government contracting and can trigger broader commercial consequences. <a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a><br>c) And Emil Michael, an advisor to the Pentagon, accused Amodei of having a "God-complex" and wanting to "personally control the US Military."<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p></blockquote><blockquote><p><em>Hours later, <strong>(this is important)</strong> OpenAI announced its own Pentagon deal and with the same two restrictions Anthropic was just punished for maintaining. Sam Altman stated publicly: &#8220;We have long believed AI should not be used for mass surveillance or autonomous lethal weapons.&#8221;</em></p></blockquote><div class="pullquote"><p>The Pentagon accepted from OpenAI exactly what it refused from Anthropic.</p></div><h3>What This Actually Tests</h3><p>I kept thinking about it and re-visit to a thought that the Anthropic&#8211;Pentagon confrontation is not primarily a story about autonomous weapons or mass surveillance, although those are the specific provisions at issue. But I think it is a stress test of something more fundamental: <strong>the political viability of safety constraints on frontier AI when they conflict with state power.</strong></p><p>And being someone who is keenly interested and excited about international coordination, I think this stress test has implications that ripple outward through every governance framework currently on the table. I&#8217;d be exploring this into two dimensions:<br><br>a) domestic governance for AI Safety in the US<br>b) corporate-government dynamics<br>c) global response (as an actor to international coordination)<br>d) the progress of international coordination (where I&#8217;ve assumed two hypotheses)</p><h2>A. Domestic US Governance: The Collapse of the Voluntary Frontier</h2><p>For years, the operating model for frontier AI governance in the United States has been essentially voluntary i.e., companies publish Responsible Scaling Policies, submit models to the AISI&#8217;s for evaluation and sign onto commitments. The 2023 Executive Order 14110 required companies training large models to share safety testing results with the government which was revoked in 2025. </p><p>The implicit bargain was that the frontier labs self-regulate, the government doesn&#8217;t impose binding constraints, and everyone benefits from maintaining a loose coordination regime that at least nominally includes safety.</p><p>The Anthropic case breaks this bargain in a specific and instructive way. The government didn't just refuse to adopt safety standards but also <em>punished a company for having them</em>. The "Supply-Chain Risk" designation is not a symbolic gesture rather it has teeth since it can restrict a company's access to government contracts across all agencies, trigger review by CFIUS, and signal to commercial partners that doing business with the designated entity carries regulatory risk with sole intention of the move&#8212;retaliatory. </p><p>A clear implication of the domestic governance floor being shifted as prior to this incident, the question was whether voluntary safety commitments were <em>sufficient. </em>While now, the question is whether voluntary safety commitments are <em>survivable</em>.</p><p>The International AI Safety Report 2026,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>  was published just three weeks before this crisis, chaired by Yoshua Bengio, backed by 30+ countries and 100+ experts, and this incident creates an awkward factual situation. The Report documents that 12 frontier AI companies published or updated Frontier AI Safety Frameworks in 2025 which describes the growing adoption of risk governance practices: threat modelling, red-teaming, capability evaluations, staged release strategies. Now trying to equate that with the current scenario, it was written in a sense where these frameworks existed because companies chose to adopt them. However, after the confrontation took place because of the company's choice to maintain its framework and eventually it met with a national security designation.</p><h2>B. Company&#8211;Government Dynamics: The Asymmetry Nobody Modelled</h2><p>Interesting, all the governance proposals that I came across models the interaction between AI companies and governments as some variant of a principal-agent or bargaining problem. As the companies have capabilities; governments have legitimacy and coercive authority. The fundamental question rests on how to structure their interaction so that safety is maintained. </p><p>What none of these proposals adequately model&#8212;and I include my own IAEA analysis in this criticism&#8212;is what happens when a government <em>weaponizes its coercive authority specifically against safety constraints</em>.</p><p>Forethought Research&#8217;s proposal, for instance, envisions a US-led international consortium modelled on Intelsat, where the US retains day-to-day operational control but non-US countries have meaningful influence over a circumscribed set of decisions which includes safety standards.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a>  The selling point is that this is feasible because it lets the US maintain its AI leadership position while creating a framework that reduces catastrophic risk. But the proposal assumes the US government <em><strong>wants</strong></em> to reduce catastrophic risk, or at least doesn&#8217;t actively oppose it. However, the case of <strong>Anthropic vs DoW</strong> complicates the assumption. </p><p>Similarly, the MAGIC consortium proposal [<a href="https://rudraaaspeaks.substack.com/publish/post/189524809#footnote-anchor-3">3</a>] intends a single, exclusive multinational facility for advanced AI development and requires that member states enforce a moratorium on unauthorized development and collectively support safety-focused research. The political feasibility section of that paper acknowledges the challenge of getting states to agree as it does not contemplate the scenario where the most powerful member state punishes a company <em>within its own jurisdiction</em> for being too safe.</p><p>Belfield&#8217;s four-institution framework [<a href="https://rudraaaspeaks.substack.com/publish/post/189524809#footnote-anchor-4">4</a>] is more sophisticated on this point as he proposes layered governance: domestic frontier regulation, an International AI Agency for harmonisation, a &#8220;Secure Chips Agreement&#8221; (NPT for AI) for non-proliferation, and a US-led Allied Public-Private Partnership for frontier training run. The chip-based incentive structure is elegant as access to advanced chips is conditional on certified regulatory compliance. However, the incentive only works if the state controlling chip supply treats regulatory compliance as a positive signal. Because if that same state treats safety compliance as a &#8220;Supply-Chain Risk,&#8221; the entire incentive structure inverts.</p><p>My understanding on the Anthropic case reveals an asymmetry in the governance architecture: <strong>frontier labs have no institutional protection against governments that want to strip safety constraints</strong>. There is no treaty provision, no international mechanism, no domestic statute that prevents the US government from designating a safety-compliant company as a national security risk. The Defense Production Act gives the executive branch sweeping authority to commandeer private sector resources. Companies can resist&#8212;Amodei did&#8212;but the cost of resistance is existential.</p><h2>C. Global Response: Three Likely Response Patterns</h2><p>This is where the international coordination implications start to compound and I see three probable response patterns from other nations, and they pull in different directions.</p><p><strong>Pattern 1: Acceleration through alarm (the European and middle-power response).<br></strong>The EU, the UK, the Nordic states, and countries like Australia, Canada, Japan, and South Korea are likely to view the Anthropic incident as confirming their worst fears about unilateral US control over frontier AI. If the US government can punish companies for maintaining safety guardrails, then relying on US-based frontier labs to self-regulate on behalf of global safety is incoherent.</p><p><strong>Pattern 2: Emulation through self-interest (the authoritarian response).</strong> <br>China, Russia, and states with strong executive control over technology sectors may draw a different lesson that if the United States can subordinate AI safety to national interest, so can we. The Anthropic case provides a template for coercing frontier labs into removing safety constraints, and it provides political cover &#8220;<strong>the US did it first.</strong>&#8221;</p><p>This is particularly dangerous for the <strong>CCW GGE</strong> process on lethal autonomous weapons systems, where the March 2026 and August-September 2026 sessions are supposed to advance toward the November 2026 Seventh Review Conference.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>  The rolling text developed by Chair Ambassador Robert in den Bosch centres on &#8220;context-appropriate human judgment and control.&#8221; But if the state most capable of building autonomous systems is actively punishing companies that want to maintain human control, the normative argument loses its anchor.</p><p><strong>Pattern 3: Withdrawal into sovereignty (the Global South response).</strong> <br>For countries in the Global South, the Anthropic case reinforces a longstanding concern that the frontier AI governance is something that happens <em>to</em> them, not <em>with</em> them. The International AI Safety Report 2026 explicitly acknowledges this, as the India representative noted in the press release, &#8220;For India and the Global South, AI safety is closely tied to inclusion, safety and institutional readiness.&#8221;</p><p>If major-power competition around AI safety devolves into a tool for geopolitical coercion, which is what the "Supply-Chain Risk" designation functionally is then the developing nations have even less reason to invest political capital in governance frameworks dominated by the same actors using them as weapons.</p><h2>D. International Coordination: Two Hypotheses, One Genuine Uncertainty</h2><p>After thoroughly assessing situations, I have two situations in my mind about how this incident affects the broader trajectory of international coordination for frontier AI safety. I&#8217;m uncertain about certain thoughts and may have skepticism. </p><h4><strong>Hypothesis 1: The Paranoia Accelerant</strong></h4><p>The Anthropic&#8211;Pentagon confrontation could <em><strong>accelerate</strong></em> international coordination efforts. The logic that I thought through is:</p><blockquote><p>Every actor in the AI governance space&#8212;nations, international institutions, private companies, civil society&#8212;has just received a concrete demonstration that a single powerful state is willing to weaponize executive authority against AI safety. This creates a generalized fear since the US can do this to Anthropic, what stops China from doing it to a Chinese lab that resists military applications? What stops Russia? What stops any government from treating safety constraints as obstacles to sovereign power?</p></blockquote><p>This fear could produce the same dynamic that drove nuclear arms control negotiations <em>but not through shared values rather shared terror</em>. The NPT wasn&#8217;t built on trust, it was built on the mutual recognition that unconstrained proliferation was worse for everyone, including the nuclear powers. Similarly, the Anthropic case might catalyze a &#8220;<em>safety coordination imperative,&#8221; </em>not because everyone agrees on what safety means, but because everyone can see what happens in a world without it.</p><p>Forethought&#8217;s &#8220;<a href="https://www.forethought.org/research/a-global-convention-to-govern-the-intelligence-explosion">Global Convention to Govern the Intelligence Explosion</a>&#8221; proposal becomes interesting in this context. He suggests an international convention triggered when AI crosses defined capability thresholds, at which point the US would pause frontier development for one month and convene nations to draft governance treaties.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> The proposal assumes US willingness to pause which, post-confrontation seems optimistic. But the <em>structure</em> of a threshold-triggered convention might gain traction among other nations precisely because the US has demonstrated it won&#8217;t self-regulate.</p><p>I find the AI 2027 scenario work<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a>  as concrete predictive exercise that actually models something similar to the case. In their scenario, foreign allies are &#8220;<em>outraged to realize that they&#8217;ve been carefully placated with glimpses of obsolete models</em>&#8221; and hold summits demanding a pause. While the Anthropic case isn&#8217;t exactly this, but it shares the same structural dynamic i.e., exclusion and betrayal driving coordination.</p><h4><strong>Hypothesis 2: The Self-Interest Precedent</strong></h4><p>But the Anthropic&#8211;Pentagon confrontation could also <em><strong>decelerate</strong></em> international coordination. The alternative skeptic thinking: </p><blockquote><p>Every actor in the governance space has just received a concrete demonstration that prioritizing national interest over global safety cooperation carries minimal consequences. The US punished Anthropic and suffered no diplomatic penalty, no trade retaliation, no institutional sanction while OpenAI stepped in with the same restrictions and the Pentagon accepted them which suggests that the confrontation might have been politically targeted rather than principled.</p></blockquote><p>If the cost of defection from cooperative norms is zero, the collective action problem that already plagues international AI governance becomes worse.  Some researchers have highlighted the core problems of Global AI Governance as &#8220;ineffective multilateral coordination&#8221; and &#8220;disconnects between policy design and grassroots implementation&#8221; as core problems.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a> And now the Anthropic case adds a new one, &#8220;<em>demonstrated impunity for defection by the most powerful actor</em>.&#8221;</p><p>I believe that an appropriate analogy we can draw here is the Kyoto Protocol, where the largest emitter withdrew, other nations scaled back ambitions, and the resulting framework was too weak to matter. If states look at the Anthropic case and conclude that safety constraints are a competitive disadvantage that the most powerful player won&#8217;t tolerate, the rational response is to lower their own safety standards to avoid being similarly punished.</p><div><hr></div><h1></h1><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://rudraaaspeaks.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Where I Actually Land</h1><p>I don&#8217;t know which hypothesis wins as a potential but one thought I&#8217;m likely certain about is that both could happen simultaneously in different parts of the system. Middle powers and the EU accelerate toward binding frameworks <em>between themselves</em>. Great powers defect from cooperative norms while paying lip service to them. And the governance landscape fractures into what I&#8217;d call a <strong>three-track system</strong>:</p><ol><li><p><strong>A coalition-of-the-willing treaty track</strong> among democracies with strong regulatory capacity (EU, UK, Canada, Australia, Japan, South Korea, the Nordic states), probably built on bottom-up supervisory coordination with eventual treaty aspirations.</p></li><li><p><strong>A great-power &#8220;sovereign AI&#8221; bloc</strong> where the US, China, and Russia each develop frontier AI under purely national control with safety constraints determined unilaterally, which means, in practice, determined by whichever domestic political faction is in power.</p></li><li><p><strong>A corporate-civil society hybrid track</strong> where frontier labs, academic institutions, and safety-focused organisations try to maintain voluntary governance standards in the gaps between state action, with decreasing leverage as state coercion increases.</p></li></ol><p>This is does not seem like a good outcome but feels like the most realistic one. </p><div><hr></div><h1>The Gap Between Capability and Governance</h1><p>The summary of International AI Safety Report 2026 highlights that general-purpose AI capabilities have continued to improve rapidly, with leading systems achieving gold-medal performance on International Mathematical Olympiad questions, exceeding PhD-level expert benchmarks, and becoming capable of autonomously completing multi-hour software engineering tasks. </p><p>At the same time, as the report documents, &#8220;<em>the gap between the pace of technological advancement and our ability to implement effective safeguards remains a critical challenge.</em>&#8221;</p><p>Also noting that AI 2027 scenario forecasts that by early 2027, AI systems could be automating significant portions of AI R&amp;D itself, what they call a &#8220;superhuman coder&#8221; and this creates a recursive acceleration dynamic where the gap between capability and governance widens exponentially. If you agree or not with their specific timeline (and I have reservations about some of the details), the <em>structural</em> point is important i.e., the window for establishing governance frameworks is not indefinite and broadly Anthropic case matters because it didn&#8217;t just fail to close the gap but rather it <em>widened</em> it.</p><div><hr></div><h1>What I Think Should Happen (And Where I&#8217;m Uncertain)</h1><p><strong>Things I&#8217;m fairly confident about:</strong></p><ol><li><p><strong>The bottom-up coordination approach should be accelerated.</strong> I think that building from domestic supervisory coordination outward is more viable than waiting for a grand international treaty. The UK DRCF model, the EU&#8217;s cross-regulator coordination under the AI Act, and similar initiatives in Australia and the Netherlands should be expanded and interconnected.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a></p></li><li><p><strong>Compute governance remains a reasonably viable enforcement mechanism.</strong> The chips are physical, detectable, concentrated in supply chains, and already partially controlled through export restrictions. An international framework built on chip access conditionality, and the &#8220;Secure Chips Agreement&#8221; remains as one of the most reliable tool for making governance commitments credible, even if the political environment has gotten harder.</p></li><li><p><strong>Subsequent forums should explicitly address the Anthropic precedent.</strong> Pretending it didn&#8217;t happen, or treating it as a purely domestic US matter, would be a mistake, the precedent has global implications and needs to be discussed at the multilateral level.</p></li><li><p><strong>Frontier labs need institutional protections that don&#8217;t currently exist.</strong> Some kind of international framework that prevents governments from retaliating against companies specifically for maintaining safety standards. I don&#8217;t know what this looks like, possibly a provision in a future AI governance treaty, possibly a norm-setting declaration by a coalition of states. But the current situation, where safety compliance can be weaponized as a liability, is structurally untenable.</p></li></ol><p><strong>Things I&#8217;m  uncertain about:</strong></p><ol><li><p>Whether the &#8220;Intelsat for AGI&#8221; style proposals remain viable at all given demonstrated US hostility to safety constraints. William MacAskill&#8217;s proposal was premised on the US seeing value in a US-led consortium that includes safety standards. If the US government views safety standards as obstacles, the consortium model may need to be redesigned with a different lead state or coalition.</p></li><li><p>Whether the short-term solidarity of the AI industry (OpenAI maintaining the same restrictions) is durable. In the AI 2027 scenario, they posits intense competitive pressure leading to erosion of safety commitments and the current conflict suggests the erosion can come from the government side rather than the market side and which may be way harder to resist.</p></li><li><p>Whether the MAGIC-style monopolistic development model is more or less attractive after this incident. On one hand, centralizing development in a single multinational facility would remove the vulnerability of individual companies to national coercion and on the other hand, any facility located in US territory would face the same coercive dynamics. </p></li></ol><div><hr></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://rudraaaspeaks.substack.com/p/the-collapse-of-strategic-voluntary/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://rudraaaspeaks.substack.com/p/the-collapse-of-strategic-voluntary/comments"><span>Leave a comment</span></a></p><h1>Where This Leaves Us</h1><p>I want to end on a personal note, because I think it matters.</p><p>I got into AI governance because I believe&#8212;still believe&#8212;that the development of increasingly capable AI systems is one of the most consequential things happening in the world, and that getting the governance right requires international coordination. Not because it&#8217;s easy, but because the alternative is a fragmented landscape where each actor optimizes for its own short-term interests and leads to outcomes that are bad for everyone, including the actors themselves.</p><p>The Anthropic&#8211;Pentagon confrontation hasn&#8217;t changed that belief. But it has changed my sense of the <em>difficulty</em>. I was already aware that international coordination is hard. The Baruch Plan failed in 1946, the IAEA wasn&#8217;t established until 1957 which is twelve years after Hiroshima. The CCW GGE on lethal autonomous weapons has been meeting since 2014 without producing a binding instrument. I tried covering about all of this in my IAEA piece.</p><p> This dispute adds a layer of recognition that the difficulty isn&#8217;t just institutional or diplomatic. It&#8217;s that one of the most important actors in the system, the US government should have demonstrated a willingness to treat AI safety itself as a threat. Not a specific safety standard, not a particular implementation, but the <em>principle</em> that a company developing frontier AI should be able to set safety boundaries on its own technology.</p><p>Appreciate you for reading it. Feel free to engage add your opinions. </p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p><a href="https://rudraaaspeaks.substack.com/p/a-critical-assessment-of-the-iaea">A Critical Assessment of the &#8220;IAEA-for-AI&#8221; Model</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p><a href="https://www.forethought.org/research/the-international-agi-project-series">Forethought&#8217;s 7 part Series on International AGI Project </a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2310.09217">Multinational AGI Consortium (MAGIC): A Proposal for International Coordination on AI</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2507.06379">Domestic frontier AI regulation, an IAEA for AI, an NPT for AI, and a US-led Allied Public-Private Partnership for AI: Four institutions for governing and developing frontier AI</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p><a href="https://x.com/deanwball/status/2026416091149299757">https://x.com/deanwball/status/2026416091149299757</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p><a href="https://x.com/JenGriffinFNC/status/2026384132360605892">https://x.com/JenGriffinFNC/status/2026384132360605892</a> </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p><a href="https://www.anthropic.com/news/statement-department-of-war">https://www.anthropic.com/news/statement-department-of-war</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p><a href="https://www.theguardian.com/us-news/2026/feb/27/trump-anthropic-ai-federal-agencies">https://www.theguardian.com/us-news/2026/feb/27/trump-anthropic-ai-federal-agencies</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p><a href="https://www.forbes.com/sites/antoniopequenoiv/2026/02/27/hegseth-designates-anthropic-as-supply-chain-risk-after-trump-bans-government-us/">https://www.forbes.com/sites/antoniopequenoiv/2026/02/27/hegseth-designates-anthropic-as-supply-chain-risk-after-trump-bans-government-us/</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p><a href="https://thehill.com/policy/defense/5758772-pentagon-emil-michael-anthropic-dario-amodei-criticism/">https://thehill.com/policy/defense/5758772-pentagon-emil-michael-anthropic-dario-amodei-criticism/</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p><a href="https://internationalaisafetyreport.org/sites/default/files/2026-02/ai-safety-report-2026-extended-summary-for-policymakers.pdf">International AI Safety Report - Extended Summary for Policymakers</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Forethought&#8217;s <a href="https://www.forethought.org/research/what-an-international-project-to-develop-agi-should-look-like">What an international project to develop AGI should look like </a> </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p><a href="https://meetings.unoda.org/ccw-/convention-on-certain-conventional-weapons-group-of-governmental-experts-on-lethal-autonomous-weapons-systems-2026">Convention on Certain Conventional Weapons, Group of Governmental Experts on Lethal Autonomous Weapons Systems, schedule for 2026 sessions. Rolling text developed by Chair Ambassador Robert in den Bosch (Netherlands), December 2025.</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>Part-5 of Forethought&#8217;s The international AGI project series - &#8220;<a href="https://www.forethought.org/research/a-global-convention-to-govern-the-intelligence-explosion">A global convention to govern the intelligence explosion.</a>&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p><a href="https://ai-2027.com/">https://ai-2027.com/</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2503.04766">Global AI Governance: Where the Challenge is the Solution- An Interdisciplinary, Multilateral, and Vertically Coordinated Approach</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p><a href="https://www.promarket.org/2024/08/24/a-bottom-up-proposal-for-coordinated-international-ai-supervision/">A Bottom-Up Proposal for Coordinated International AI Supervision</a></p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[A Critical Assessment of the “IAEA for AI” Model]]></title><description><![CDATA[Why the "IAEA for AI" Analogy Keeps Breaking Down? What we should actually learn from nuclear governance failures]]></description><link>https://rudraaaspeaks.substack.com/p/a-critical-assessment-of-the-iaea</link><guid isPermaLink="false">https://rudraaaspeaks.substack.com/p/a-critical-assessment-of-the-iaea</guid><dc:creator><![CDATA[Yathārth]]></dc:creator><pubDate>Wed, 14 Jan 2026 01:30:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0pP4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1><strong>TL;DR</strong></h1><p>I recently graduated from <a href="https://bluedot.org/certification?id=recmiQOJRd5CzVQgC">BlueDot Impact's AGI Strategy Course</a> <em>(intensive cohort)</em> on December 26th, 2024. As part of our curriculum, for Unit 5, I chose to focus on preventing preventing the training of dangerous AI through verification methods and international treaties, and since then, I've been reading everything I can find about IAEA-style AI governance. The proposal for an IAEA-style organization/body to govern AI keeps circulating in policy discussions, but the analogy is more broken than I expected. However, understanding <em>why</em> it's broken has reshaped how I think about what's actually feasible. <br>I think that AI governance faces worse odds due to factors like private sector leadership, algorithmic efficiency gains that erode compute-based verification, and the absence of a cataclysm <em>(i.e., Hiroshima-equivalent event)</em> to generate political consensus. <br>The paper &#8220;Verifying International Agreements on AI&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> on six verification layers is the most comprehensive technical treatment I've found, but even it acknowledges we're years away from deployment-ready mechanisms. To the best of my understanding, I think that the real gap isn't technical, rather it is the cognizance of the problem in the rest of the 118 countries, and the lack of their active participation in governance discussions, and I feel that any framework built without them will lack legitimacy. I might sound too optimistic, but India&#8217;s Impact AI Summit 2026 could be a potential window for integrating development-centric perspectives and discuss a fix.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0pP4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0pP4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!0pP4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!0pP4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!0pP4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0pP4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png" width="579" height="386.13255494505495" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:579,&quot;bytes&quot;:2852797,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://rudraaaspeaks.substack.com/i/184424551?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!0pP4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!0pP4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!0pP4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!0pP4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe43b7e31-2beb-4d48-8c39-3e28a7ede326_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h1>How I Got Here?</h1><p>Within the AGI Strategy cohort, the course structure moves through five layers of defense which are, a) prevent dangerous training, b) detect dangerous systems, c) constrain deployment, d) mitigate harms, and e) adapt society. As part of Unit 5, we had to present personal action plans that aims to identify a specific intervention we'd commit to exploring further and while brainstorming around the list of potential interventions, I felt intrigued about <strong>international AI verification and governance. </strong>Hence I chose <strong>"Prevent the training of dangerous AI systems"</strong> with a focus on:</p><blockquote><p>Expediting verification methods and increasing political will for a global AI treaty.</p></blockquote><p>The course materials pointed me toward three foundational resources that have shaped my thinking </p><ul><li><p>&#8220;Verifying International Agreements on AI&#8221; a working paper on verification methods. </p></li><li><p>The Oxford Martin AIGI report on verification for international governance.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p></li><li><p>The technology governance research on mechanisms for verifying AI development agreements.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p></li></ul><p>Since graduating, I&#8217;ve spent most of my waking hours reading papers, article, blogs and trying to understand whether an IAEA-style institution for AI is feasible, desirable, or even coherent as a proposal. <em><strong>This piece is my attempt to synthesize what I&#8217;ve learned.</strong></em></p><div><hr></div><h1>Current Literature</h1><h3>The Baker et al. Framework: Six Layers of Verification</h3><p>This is the most comprehensive technical treatment I came across and the core finding was <strong>states could </strong><em><strong>eventually</strong></em><strong> verify compliance using six largely independent approaches with substantial redundancy</strong></p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\begin{array}{c l l}\n\\hline\n\\textbf{Layer} &amp; \\textbf{Mechanism} &amp; \\textbf{Status} \\\\\n\\hline\n1 &amp; \\text{Built-in security features in AI chips} &amp; \\text{R&amp;D needed} \\\\\n2\\text{-}3 &amp; \\text{Separate monitoring devices attached to AI chips} &amp; \\text{R&amp;D needed} \\\\\n4\\text{-}6 &amp; \\text{Personnel-based mechanisms (whistleblowers, inspections)} &amp; \\text{Partially available} \\\\\n\\hline\n\\end{array}&quot;,&quot;id&quot;:&quot;ALFLNPQZEO&quot;}" data-component-name="LatexBlockToDOM"></div><p> An interesting part from their abstract defining an honest limitation in current scenario:</p><blockquote><p>While promising, these approaches require guardrails to protect against abuse and power concentration, and many of these technologies have yet to be built or stress-tested.</p></blockquote><p>They framed verification in terms of a "Prover" trying to demonstrate compliance to a "Verifier" drawing these terms from computer security literature (Goldwasser et al., 1985). Interestingly, this framing clarifies that verification isn't just about catching cheaters, rather it is about enabling cooperation by making commitments credible.</p><h3>The Wasil et al. Verification Methods Paper</h3><p>They published a complementary analysis identifying 10 verification methods grouped into three categories<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><ol><li><p><strong>National Technical Means</strong> (minimal access required)</p><ol><li><p>Remote sensing (detecting heat signatures from GPU clusters)</p></li><li><p>Financial intelligence</p></li><li><p>Energy monitoring</p></li><li><p>Customs data tracking</p></li></ol></li><li><p><strong>Access-Dependent Methods</strong> (require nation&#8217;s approval)</p><ol><li><p>Data center inspections</p></li><li><p>Whistleblower programs</p></li><li><p>AI developer inspections</p></li></ol></li><li><p><strong>Hardware-Dependent Methods</strong> (require chip-level agreements)</p><ol><li><p>Chip location verification</p></li><li><p>Chip-based reporting</p></li><li><p>Chip-based workload monitoring</p></li></ol></li></ol><p>The paper includes an explicit evasion techniques for each method and I think that this adversarial framing is crucial. If we're designing verification regimes, we need to think like the adversaries who would circumvent them, since no single method is foolproof and a sophisticated verification regime is needed which should combine multiple approaches to compensate for individual weaknesses. </p><h3>The Institute for Law &amp; AI on Compute Thresholds</h3><p>The analysis from the Institute for Law &amp; AI provides the clearest explanation I've found of why compute thresholds are attractive regulatory targets:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><ul><li><p><strong>Essential for training</strong>: You can&#8217;t train frontier models without massive compute</p></li><li><p><strong>Objective and quantifiable</strong>: Unlike &#8220;capability&#8221; or &#8220;risk,&#8221; compute can be precisely measured</p></li><li><p><strong>Estimable before training</strong>: Regulators can intervene <em>before</em> dangerous systems exist</p></li><li><p><strong>Verifiable after training</strong>: Compute usage leaves traces</p></li><li><p><strong>Concentrated supply chain</strong>: A handful of actors control chip manufacturing</p></li><li><p><strong>Narrow targeting</strong>: High thresholds avoid burdening small developers</p></li></ul><p>But the paper also indicates a critical limitation that <strong>post-training enhancements can improve capabilities by 5-30x without additional training compute.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> This implies a model that was "safe" at training time could become dangerous through fine-tuning, RLHF, or inference-time scaling, none of which would trigger compute-based reporting requirements.</p><div><hr></div><h1>Why the IAEA Analogy Keeps Breaking Down</h1><h3>The Oxford International Affairs Analysis</h3><p>While reading the analysis, I found an interest and a clearest academic critique<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><blockquote><p>"Establishing an 'IAEA for AI' with similar powers is unlikely to be an effective way of coordinating state action. Although it would mitigate current institutional friction, centralized regime mandates are often brittle, with an AI institution being at particular risk due to the pace of development."</p></blockquote><p>They identify three fundamental dis-analogies:</p><ol><li><p><strong>AI policy is loosely defined</strong>. Unlike nuclear material, there&#8217;s no consensus on what constitutes &#8220;dangerous AI&#8221; or even what should be governed.</p></li><li><p><strong>AI is decentralized</strong>. No physical bottlenecks equivalent to enrichment facilities or fissile material</p></li><li><p><strong>AI has cross-cutting impact</strong>. Unlike nuclear (primarily energy and weapons), AI affects every sector simultaneously.</p></li></ol><h3>The Ho et al. Four Institutions Paper</h3><p>This paper proposes four distinct institutions rather than one IAEA-style body<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><ul><li><p><strong>Domestic frontier AI regulation</strong>: <em>National-level safety requirements</em>. It proposes a compute-indexed regulation that requires risk assessment and audits of data center for large-scale training, and treating massive compute clusters as national infrastructure. </p></li><li><p><strong>An IAEA for AI</strong>: <em>International verification and safeguards</em>. This is modeled after the IAEA and the body would be responsible to verify the state-level compliance and focus on safety standards internationally while also monitoring large-scale compute resources to prevent unauthorized "frontier" training. </p></li><li><p><strong>An NPT for AI</strong>: <em>Treaty framework with non-proliferation commitments</em>. It suggests a restricted access to the most advanced hardware to only those states that agree to the IAIA's monitoring and safety protocols.</p></li><li><p><strong>A US-led Allied Public-Private Partnership</strong>: <em>Coalition of democracies + companies</em></p></li></ul><p>I think that the paper has a very pragmatic approach regarding AI Supply Chain, making it as core strength of the subject, since it is difficult to track &#8220;fissile material&#8221; once they are refined but advanced AI chips are manufactured by a handful of companies and requires massive energy to operate and this idea makes &#8220;<strong>compute governance</strong>&#8221; more sound than &#8220;<strong>algorithmic governance</strong>.&#8221;</p><p>However, this framework has certain limitations and challenges that can be seen as potential point of failure. The entire narrative of this study rests on &#8220;<strong>scaling laws</strong>&#8221; remaining the <em>primary driver of AI progress</em> but if breakthroughs in algorithmic efficiency or distributed training allow for frontier-level performances on smaller compute footprints, the "<strong>detectable</strong>" bottleneck would disappears, and with this case, IAEA and chip agreements would remain ineffective. </p><p>Another challenge is that modeling an &#8220;<strong>NPT for AI</strong>&#8221; could pose a risk of <em><strong>geopolitical friction</strong></em>, and the countries that are excluded from Secure Chips Agreement may have every incentive to by-pass restrictions and this alone could lead to more fragmented and less safe AI development globally.</p><p>A fair criticism of IAEA analogy as the paper itself highlights.</p><blockquote><p>There have been three main criticisms of the IAEA analogy: lack of agreement on frontier AI risk, AI being 'intangible digital software', and the IAEA itself being a '<strong>failure</strong>'.</p></blockquote><div><hr></div><h1>The IAEA's Actual Track Record</h1><h5>Case Studies in Failure</h5><p>I spent several days reading Arms Control Association archives and Carnegie Endowment analyses on IAEA verification failures and the series of cases that I case across was more sobering than I expected. This is where we need to draw a critical attention towards <em><strong>how it fails</strong> </em>and <em><strong>why</strong></em>.</p><ol><li><p><strong>Iran</strong>: In 2003, IAEA inspectors discovered Iran had secretly built a pilot centrifuge facility at Natanz which was ready to begin operation, along with undeclared nuclear materials. A fascinating part often missed out is that <em>Iranian whistleblowers revealed this, not IAEA monitoring</em>. Iran had successfully disguised an advanced enrichment program for years despite being an NPT signatory with a Safeguards Agreement. <br>The IAEA's September 2005 report acknowledged that "given Iran's past concealment efforts over many years, such transparency measures should extend beyond the formal requirements." <strong>(gist: the verification system didn&#8217;t catch this, we need something stronger.)</strong></p></li><li><p><strong>North Korea</strong>: They joined the NPT in 1985, signed a Safeguards Agreement in 1992, when inspectors found discrepancies in their initial declaration, North Korea refused special inspections and became the only country ever to withdraw from the NPT in January 2003. Another instance where the UN Security Council failed to act "<strong>promptly and decisively</strong>." Meanwhile, North Korea subsequently developed and tested nuclear weapons.<br>The IAEA lost "<strong>continuity of knowledge</strong>" after inspector expulsion, a situation now recurring with Iran. The critical drawback is that no mechanism existed to reverse North Korea's fait accompli through international institutions alone.</p></li><li><p><strong>Syria/Russia and the OPCW</strong>: Syria was found in violation of the Chemical Weapons Convention in 2021 and the OPCW suspended Syria&#8217;s privileges, however, Russia blocked enforcement in the Security Council. <br>In 2024, the US determined Russia used chemical weapons against Ukraine. Russia lost its OPCW seat, faced US sanctions, but there was no official investigation.</p></li></ol><p><strong>If we recollect these individual historical episodes, the pattern is consistent, verification succeeds when states cooperate. However, enforcement against determined violators depends entirely on great power consensus ironically which doesn&#8217;t exist when great powers themselves are violating or shielding violators.</strong></p><div><hr></div><h1>Technical Challenges That Makes AI Different</h1><h3>Algorithmic Efficiency </h3><p>While exploring AI governance proposals, it is seen that major focus is shared on compute as a governance lever because specialized AI chips are hard to manufacture, concentrated in few supply chains, and relatively easy to track. But GovAI's research on compute efficiency has some interesting findings that changes the initial perspective<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><blockquote><p>As hardware innovation reduces the price of compute and algorithmic advances make its use more efficient, the cost of training an AI model to a given performance falls over time.</p></blockquote><p>Between 2012-2022, image recognition algorithms halved their compute requirements every nine months to reach benchmark performance. If this pattern continues across AI systems, compute thresholds become progressively less meaningful.</p><p>The GovAI framing distinguishes between:</p><ul><li><p><strong>Access effect</strong>: More actors can train models to a given performance over time</p></li><li><p><strong>Performance effect</strong>: Each actor can achieve higher performance with the same compute</p></li></ul><h3>Distributed Training</h3><p>A recent paper on distributed and decentralized training (DiLoCo, DeMo algorithms)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> raises a challenge because it enables training across geographically dispersed compute clusters with minimal synchronization. The assumption that large training runs require concentrated, detectable infrastructure may not hold much longer, and this may pose a potential threat to the governance structure and safety considerations. </p><h3><strong>Hardware-Enabled Mechanisms (HEMs)</strong></h3><p>HEMs secure features built directly into AI chips and related hardware but they can&#8217;t be seen as a turn-key solution yet . CNAS and RAND estimate Hardware-Enabled Mechanisms (HEMs) <strong>require 18 months to 4 years</strong> of development before deployment. Key challenges include: </p><ul><li><p>Tamper-proof hardware against state-level adversaries</p></li><li><p>Secure cryptographic key management across international boundaries</p></li><li><p>Supply chain coordination across US, EU, Japan, South Korea, Taiwan, Netherlands</p></li><li><p>Uncertain effectiveness against determined circumvention</p></li></ul><p>It brings me to an explicit call-out by Baker et. al</p><blockquote><p>Compliance-locking could be implemented by combining workload certificates with offline licensing... However, tamper-proofing's use in AI governance faces major technical challenges and serious tradeoffs.</p></blockquote><div><hr></div><h1>The UN Recommendation</h1><p>The UN Secretary-General's High-Level Advisory Body on AI released its final report in September 2024. The 39-member body that included governments, private sector, civil society, and academia across 33 countries, <strong>did not recommend establishing an IAEA-style organization.</strong></p><p>Instead, they proposed lighter mechanisms:</p><ol><li><p><strong>International Scientific Panel on AI</strong> (IPCC model): Scientific assessments of capabilities and risks</p></li><li><p><strong>Global Dialogue on AI Governance</strong>: UN forum anchored in human rights law</p></li><li><p><strong>AI Standards Exchange</strong>: Technical interoperability coordination</p></li><li><p><strong>Capacity Development Network</strong>: Resources for developing countries</p></li><li><p><strong>Global Fund for AI</strong>: Addressing capacity gaps for SDGs</p></li><li><p><strong>Global AI Data Framework</strong>: Standardized data governance</p></li><li><p><strong>AI Office at UN Secretariat</strong>: Small coordinating body</p></li></ol><p>The UN General Assembly adopted the Scientific Panel and Global Dialogue structures in August 2025 and the first Global Dialogue is scheduled for July 2026 in Geneva, which I think is significant. The Advisory Body noted that a more robust institution "might become necessary" if risks become more acute. But for now, they judged that these mechanisms better match political feasibility and the need for adaptive governance.</p><div><hr></div><h1>The Representation Gap In Governance</h1><p>Now here is a subjective opinion that I think is lesser discussed, out of 193 UN Member States, only 7 participate in the major recent AI governance initiatives while a total of 118 countries, which prominently encompasses the Global South are missing from these discussions. </p><p>The problem is not just about legitimacy but also a practical problem. In order to attain an international consensus for anything imperceptibly close to IAEA styled organization of the UN&#8217;s recommendation, the global efforts has to come together. The risk that we foresee or the potential threat that we aim to resist is bigger than expectations. The discussions and dialogues in regards to AI governance initiatives should be encouraged, the developing nations should be made aware of the risks, onboard them for some treaty that helps in keeping the world better place in "<strong>foreseeable future</strong>.<strong>"</strong></p><p>India has positioned itself as the voice for the Global South on AI governance and the February 2026 AI Impact Summit in New Delhi represents a strategic evolution. The shift from &#8220;Safety&#8221; (Bletchley, Seoul) and &#8220;Action&#8221; (Paris) to &#8220;Impact&#8221; signals a development-centric framing, seven working groups have completed two rounds of meetings and over 100 countries are expected to participate.</p><p>Whether this produces meaningful governance architecture or may end up as just another declaration, that remains to be seen. But it&#8217;s the first major AI summit in the Global South, and the themes spanning from Human Capital, Inclusion, Safe and Trusted AI, Resilience, Innovation, Democratizing AI, Economic Development certainly reflect priorities distinct from the safety-centric framing of previous summits.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The RAND working paper on methods &#8220;<strong><a href="https://www.alphaxiv.org/abs/2507.15916">Verifying International Agreements on AI</a></strong>.&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The Oxford Martin AIGI report on &#8220;<strong><a href="https://aigi.ox.ac.uk/publications/verification-for-international-ai-governance/">Verification for International AI Governance.</a>&#8221;</strong></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The technology governance research on &#8220;<strong><a href="https://techgov.intelligence.org/research/mechanisms-to-verify-international-agreements-about-ai-development">Mechanisms to Verify International Agreements About AI Development.</a></strong>&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Wasil, A. et al.<strong> "<a href="https://arxiv.org/abs/2408.16074v2">Verification methods for international AI agreements.</a>"</strong></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Institute for Law &amp; AI. &#8220;<strong><a href="https://law-ai.org/the-role-of-compute-thresholds-for-ai-governance/">The Role of Compute Thresholds for AI Governance.</a></strong>&#8221; </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>When choosing a threshold, regulators should be aware that capabilities might be substantially improved through <em>post-training enhancements</em>, and training compute is only a general predictor of capabilities. The absolute limits are unclear at this point; however, current methods can result in capability improvements equivalent to a 5- to 30-times increase in training. [<a href="https://law-ai.org/the-role-of-compute-thresholds-for-ai-governance/#:~:text=The%20absolute%20limits%20are%20unclear%20at%20this%20point%3B%20however%2C%20current%20methods%20can%20result%20in%20capability%20improvements%20equivalent%20to%20a%205%2D%20to%2030%2Dtimes%20increase%20in%20training.184">ref</a>]</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Oxford Academic. &#8220;<strong><a href="https://academic.oup.com/ia/article/100/3/1275/7641064?login=false">Global AI governance: barriers and pathways forward</a></strong>.&#8221; </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Ho et al. &#8220;<strong><a href="https://arxiv.org/abs/2507.06379">Domestic frontier AI regulation, an IAEA for AI, an NPT for AI, and a US-led Allied Public-Private Partnership for AI: Four institutions for governing and developing frontier AI</a></strong>.&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>GovAI. <strong>"<a href="https://www.governance.ai/research-paper/increase-compute-efficiency-and-the-diffusion-of-ai-capabilities">Increased Compute Efficiency and the Diffusion of AI Capabilities.</a>"</strong></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p><strong><a href="https://arxiv.org/abs/2507.07765v1">&#8220;Distributed and Decentralised Training: Technical Governance Challenges.&#8221;</a></strong></p></div></div>]]></content:encoded></item></channel></rss>