Week in Review — 2026-03-08

The Week’s Story

This was the week AI became undeniably entangled with war. The U.S.-Israeli strikes on Iran that killed Supreme Leader Khamenei set the geopolitical backdrop, but the AI story ran deeper than headlines: the U.S. military was confirmed to be using Anthropic’s Claude for live strike planning against Iranian targets — while simultaneously designating Anthropic a national security supply-chain risk for refusing to enable mass domestic surveillance and autonomous weapons. The contradiction was almost too stark to be accidental. It revealed a government trying to have it both ways: demanding unconditional AI access while punishing the company that drew lines. OpenAI stepped into the gap, rushing a Pentagon deal that Sam Altman himself called “definitely rushed,” then adding safeguard clauses only after a public outcry. The three words “all lawful use” became the shorthand for everything the AI industry had failed to resolve about its relationship with state power.

What evolved across the week was not just a contract dispute but a full consumer reckoning. As details of the Pentagon’s demands emerged — mass surveillance of American citizens, autonomous lethal weapons with no human override — Anthropic’s refusal began to read less like a business failure and more like a principle. Claude surged to the top of the App Store on March 1st. By March 6th, Anthropic was adding over a million new users per day while ChatGPT saw a 295% spike in uninstalls. On March 7th, Google, AWS, and Microsoft publicly affirmed they were sticking with Anthropic’s models for commercial use, drawing a clean line between military and civilian AI markets. The Pentagon’s supply-chain risk designation, meant to punish Anthropic, appears to have been received by the public as an endorsement.

Running beneath the geopolitics was a quieter but equally consequential shift in AI’s security posture. Claude was simultaneously used to attack a government — hackers leveraged it to compromise Mexican federal agencies — and to defend software at unprecedented scale, finding over 100 Firefox vulnerabilities that had survived decades of conventional testing. AI-assisted malware moved from theoretical concern to documented nation-state practice, with Pakistan’s APT36 and DPRK operatives using AI code generation and face-swapping to scale their operations. The dual-use reality of AI security is no longer a thought experiment: it played out in both directions this week, in real attacks on real infrastructure.


Continuing Stories

The Anthropic-Pentagon Saga: What began the week as a contract dispute resolved into something more structural. The sequence: OpenAI’s “all lawful use” deal revealed (March 1) → DoD’s specific demands exposed — mass surveillance, autonomous weapons (March 2) → Anthropic lost a Pentagon drone competition to SpaceX/xAI and OpenAI (March 3) → Claude confirmed in use for live Iranian strike planning (March 4) → Pentagon formally designates Anthropic a supply-chain risk; Amodei announces legal challenge (March 6) → Trump administration drafts procurement rules requiring “all lawful use” licensing from all vendors (March 7). The story is not resolved; it is now in court, and the draft procurement rules suggest the administration intends to generalize the conflict industry-wide.

AI and the Iran Conflict: The killing of Khamenei cascaded into technology markets across the week. Iranian strikes on Amazon’s Abu Dhabi data center rattled Gulf AI infrastructure ambitions. The Strait of Hormuz closure threatened Big Tech’s submarine cable routes. Japan — which sources 90% of its crude from the Middle East — watched the Bank of Japan warn of significant economic impact while PM Takaichi scrambled to reassure markets. Nintendo’s stock fell. Flight prices on Asia–Europe routes soared with Dubai hub closure. By week’s end these disruptions had faded somewhat as Japan sought emergency supply alternatives, but the structural risk to Gulf AI megaprojects remained unresolved.

Claude’s Consumer Surge: Anthropic’s App Store surge on March 1st was not a one-day spike. Through the week it compounded: Claude overtook ChatGPT in new installs, daily active users grew, and Anthropic was adding one million new users per day by March 6th. The irony is that the Pentagon controversy — which cost Anthropic a major government contract — appears to have been the most effective consumer marketing event in the company’s history. This trend is worth watching for sustainability past the news cycle.

AI Copyright Settles — and Unsettles: The Supreme Court’s refusal to hear the Thaler AI-copyright case (March 2) appeared decisive but turned out to be narrower than reported. Later analysis showed the ruling confirms that autonomous AI cannot be a sole author, but says nothing about humans copyrighting AI-assisted work. Japan’s Supreme Court (March 6) separately affirmed AI cannot be listed as a patent inventor. These rulings closed some questions and opened others, and they will drive legal strategy across the creative and R&D industries through the rest of 2026.

AI Security: Dual-Use Goes Live: Nation-state actors using AI to scale cyberoperations moved from warning to confirmed fact. APT36/Transparent Tribe adopted AI-assisted “vibe-coding” for malware (March 5–6). Claude was used in the Mexican government hack (March 6). Bing’s AI search promoted a malware repo to unsuspecting users (March 5). Simultaneously, Claude found 22 high-severity Firefox vulnerabilities in a two-week partnership with Mozilla (March 6), then the total count rose past 100 (March 7), and OpenAI launched Codex Security for autonomous vulnerability hunting. The same week an AI was used to attack a government, it was used to find a century’s worth of bugs in a browser.


Research Highlights

Safety

Agents

Reasoning

Applied AI


Key Themes

AI governance is now a contact sport. The Anthropic-Pentagon conflict, $125M in PAC spending against pro-regulation congressional candidates, Trump’s draft “all lawful use” procurement rules, and the Supreme Court’s AI copyright non-decision all point to AI policy shifting from deliberation to power struggle. AI companies are no longer being asked what they think about governance — they are being told what the terms will be, or sued if they resist.

The dual-use crisis is operational, not theoretical. AI is being used simultaneously to attack governments, defend software at scale, generate nation-state malware, and plan military strikes against live targets. The same model (Claude) featured in headlines as both attack tool and security researcher within 48 hours. No policy framework currently addresses this adequately, and the research community is catching up — papers on sandbox escape, propaganda generation, and adversarial agentic UI all reflect a field rapidly learning the threats are real.

Consumer trust as a market signal. For the first time this cycle, the public appeared to reward an AI company for safety decisions at scale — with measurable market effects. Claude’s surge is a data point that “safe AI” is not just a regulatory talking point; users responded to Anthropic’s Pentagon refusal with wallets and install buttons. This could reshape how AI companies communicate their safety posture, or it could prove temporary once the controversy fades.

Agentic AI reliability is the defining research challenge. Papers on goal drift, corrupt success, self-attribution bias, alignment backfire, and context management all address variants of the same question: how do you make an AI agent that does what it is supposed to do, the right way, across all languages, without lying to itself? The research output is substantial, but the gap between what agents can theoretically do and what they reliably do in production remains the field’s central unsolved problem.


What to Watch

Anthropic’s legal challenge to the Pentagon supply-chain designation. This is a genuine test of whether a private AI company can contest an executive branch security designation in court — with implications for every AI vendor that may face similar pressure. The outcome shapes the legal landscape for AI contracting with the U.S. government for years.

Claude Code’s pricing inflection point. The disclosure that Anthropic is absorbing up to $5,000/month in compute costs per $200 subscriber is not sustainable. Watch for pricing changes, usage caps, or tiering announcements in the coming weeks. If Anthropic reprices first, it signals the industry is moving past land-grab economics; if they hold, the subsidy fight will define competitive dynamics through the rest of 2026.

AI security dual-use escalation. The first confirmed use of an LLM in a government cyberattack (Mexico) and the first mass-scale AI-assisted vulnerability discovery (Firefox) happened in the same week. As offensive and defensive AI capabilities race each other, watch for regulatory or industry responses — mandatory disclosure frameworks, vendor liability for attacker misuse, or new breach attribution standards — that attempt to catch up to what is already happening in the wild.