Skip to main contentSkip to main content
AI Governance · Fiduciary Risk

The Pentagon-OpenAI Deal:Why "Safety Stacks"Are a Fiduciary Risk for Every Board

The deal replaced Anthropic's legal guardrails with OpenAI's elastic software controls. The net effect: replacing the lock on your corporate vault with a curtain.

June Lai
14 min read
CFACPACMABSc (Biochemistry)Head of AI Governance
€35M
Max EU AI Act fine
or 7% global turnover
Aug 2026
EU AI Act enforcement
High-risk systems deadline
3
Override mechanisms
Codified in the deal
5
Preventative controls
Boards must verify
TL;DR — Executive Summary

In late February 2026, the Pentagon blacklisted Anthropic for refusing to remove safety restrictions. Hours later, OpenAI struck a deal with the same restrictions — but framed differently: as a technical “safety stack” rather than legally hardcoded terms.

The deal replaced Anthropic's legal guardrails with OpenAI's software-based controls — akin to replacing the lock on your corporate vault with a curtain. A government override, a cyber-attack, or a contract reclassification can pull it back at any time.

If your company uses any cloud-based AI provider, this precedent affects you. Your data sits on servers subject to sovereign authority. Your “privacy policy” is a soft lock. The only defence is a hard lock — a technical control that you own.

Three things every director must do now:

01

Ask your CTO: "If our AI vendor is served with a national security override, what technical barrier — not legal clause — prevents our data from being accessed?"

02

Demand a kill switch: A hardware-level mechanism at your network perimeter that severs the data pipeline if safety thresholds are breached.

03

Verify your Human-in-the-Loop: Confirm whether your AI safety officer has genuine authority to veto unsafe deployments — or whether they are a rubber stamp.

01 · The Facts

What Happened: The Pentagon-OpenAI Deal

On 27 February 2026, the Trump administration designated AI company Anthropic a “supply chain risk to national security” — a penalty historically reserved for companies from adversary nations such as Huawei. Defence Secretary Pete Hegseth declared that no contractor, supplier, or partner working with the US military may conduct any commercial activity with Anthropic.

The dispute centred on two “red lines” Anthropic insisted upon: a prohibition on mass domestic surveillance of US citizens, and a prohibition on fully autonomous weapons without human involvement. The Pentagon demanded Anthropic make its AI available for “all lawful purposes” without company-imposed limitations. Anthropic refused, stating it could not “in good conscience” accede.

Hours later, OpenAI announced a deal with the Pentagon. CEO Sam Altman stated the agreement contained stronger guardrails than Anthropic's. Altman admitted the deal was “definitely rushed” and acknowledged “the optics don't look good.”

The Critical Difference

Anthropic's Approach (Rejected)

Legally hardcoded restrictions — absolute limits requiring an act of parliament to change. Constitutional-level protection.

OpenAI's Approach (Accepted)

Software “safety stack” — elastic controls that follow current US law. One memo can change them.

Critical Risk for Non-US Companies

An amendment restricts tracking of US persons — but this does not protect foreign nationals or international companies, leaving them vulnerable to “lawful ingestion” of their data under existing US surveillance laws.

02 · Why This Matters

The Governance Failure: Why This Affects Your Company

This is a material shift in AI risk for corporate directors with international operations, GDPR-regulated entities, or any company that processes data through a US cloud AI provider.

The Analogy

“No board would accept this in any other context. You would not rely on the engine to decide if the driver is speeding. You would install a speed limiter and a kill switch.”

— June Lai, Head of AI Governance

The Three Governance Flaws

01Legal

The "Lawful Purpose" Loophole

OpenAI's restrictions follow current US law. Unlike Anthropic's legally hardcoded restrictions, they are elastic. If Department of War policies change to permit AI-directed lethal force under specific combat conditions, OpenAI's contract follows that legal direction.

Director's Perspective

This is the difference between a constitutional right and an internal company policy. By omission, the contract signals that non-Americans do not have the same protections. For a national privacy regulator, this is prima facie evidence that data is not being treated with equivalent domestic security.

02Cybersecurity

The Safety Stack Is a Cybersecurity Target

By centralising ethical controls into a software layer, the Pentagon has turned AI safety into a cybersecurity problem. The safety stack is now the highest-value target in the system. A successful attack on Azure GovCloud or OpenAI's verification servers could disable safety filters entirely.

Director's Perspective

When your ethical guardrails are purely software-based, AI safety is cybersecurity. Your CISO's threat model must include the failure of your AI vendor's safety controls — not just your own.

03Independence

The Human-in-the-Loop Is Under Pressure

OpenAI's cleared engineers are embedded within the Pentagon's command structure. In practice, a private-sector engineer cannot override a military General declaring a 'cyber emergency' in a war room. These engineers hold security clearances that bind them under the Secrecy Act.

Director's Perspective

A Human-in-the-Loop who lacks authority, independence, and legal protection to intervene is not a control — they are a witness. Governance frameworks (NIST AI RMF, EU AI Act) require genuine intervention power, not observation rights.

03 · Technical Architecture

The “Safety Stack” — What It Actually Is

For directors unfamiliar with the technical architecture, here is what OpenAI's “multi-layered approach” consists of — and where each layer fails.

Cloud-Only Deployment

Governance Value

OpenAI models run in the cloud rather than on edge systems. This allows OpenAI to maintain real-time control and update the safety stack remotely. A cloud-based model can be updated, monitored, and disconnected remotely.

Governance Risk

Cloud-only deployment means all controls are centralised. A single breach of the cloud infrastructure compromises the entire safety system globally.

Real-Time Classifiers and Refusal Policies

Governance Value

Automated classifiers detect unsafe queries. If a prompt violates a red line, the stack can block the output, route it to a less capable model, or trigger a refusal. Real-time detection operates at scale.

Governance Risk

Classifiers detect obvious violations. They are far less effective at detecting statistical bias or aggregate harm — a model ranking 10,000 individuals may constitute a prohibited social credit system without any single output appearing dangerous.

Embedded Personnel

Governance Value

Cleared OpenAI safety researchers work alongside military staff to monitor model usage and verify compliance. OpenAI retains the right to modify its safety stack and have models refuse specific tasks.

Governance Risk

These personnel are positioned as witnesses (who log events) rather than authorities (who prevent them). NIST AI RMF and EU AI Act require the latter.

Termination Clause

Governance Value

If the Pentagon violates contract terms, OpenAI can terminate the agreement and revoke the licence. Contractual leverage provides a legal "nuclear option."

Governance Risk

In an active conflict, the Pentagon could invoke Defence Production Act powers to keep servers running, rendering OpenAI's termination right symbolic.

04 · Override Mechanisms

How “Red Lines” Get Bypassed

The "Dual-Key" Override

The deal includes an override protocol for "cyber emergencies." A high-ranking military officer and a cleared OpenAI engineer can manually bypass a safety trigger if deemed a false positive caused by enemy interference. The terms "cyber emergency" and "active interference" are not strictly defined in US law — the Department of War, not OpenAI, determines whether an AI refusal is legitimate or enemy spoofing.

The "Lawful Purpose" Expansion

OpenAI's restrictions follow current US law. If the law changes — through executive order, updated Department of War policy, or wartime emergency powers — the restrictions change with it. Anthropic's approach attempted to set absolute limits that would survive legal change.

The "National Security" Master Key

Under the US Defence Production Act, the government can compel US vendors (OpenAI, Microsoft, Google, Amazon) to grant access to any data on their servers for national defence. Encryption does not stop a federal subpoena. If your company's data sits on a US-hosted cloud, this is not a theoretical risk — it is a legal reality the OpenAI-Pentagon deal simply codifies.

The Corporate Implication

Your company is transferring data to a provider that has explicitly authorised military access. You are not protected by their privacy policy. You are not protected by GDPR alone. A “Schrems III” legal disaster is waiting to happen.

05 · Board Framework

Implications for Every Board: Policy to Architecture

The question is no longer “Does our AI vendor have a good privacy policy?” It is: “What technical barrier prevents our data from being accessed if our vendor's safety controls are overridden?”

Why GDPR Is No Longer Sufficient

GDPR (Forensic Shield)

Penalises you after a breach has occurred. Tracks where data lives. Cannot detect if an AI uses “consented” data for a “prohibited” purpose.

EU AI Act (Architectural Shield)

Requires tracking how data is transformed and used by AI models. Fully enforceable August 2026. Fines up to €35M or 7% of global turnover.

The Three Pillars of Preventative Governance

The Kill Switch

A real-time intervention power — a legal and technical mandate to sever a model's connection to your data if a pre-defined safety threshold is breached. Hardware-level control at your network perimeter, not a software permission inside the vendor's cloud.

Data Sovereignty Architecture

Zero-trust data pipelines that treat the AI model as an "untrusted" entity. Client-side encryption where your company holds the only keys in a hardware security module (HSM) on-premise. If the vendor is ordered to hand over data, they cannot decrypt it.

Multi-Vendor Verification

A secondary, independent AI from a different developer must co-sign high-stakes outputs. It is exponentially harder to compromise two different neural architectures simultaneously than to bypass a single dual-key override.

Board Readiness Assessment

Board Readiness Assessment: risk areas, questions for management, and 2026 control requirements
Risk AreaBoard Question2026 Control Requirement
Data SovereigntyDoes the board have a physical kill switch to sever the data pipeline if the AI vendor is compromised or overridden?Hardware-level isolation: cryptographic gating that prevents data access even if the vendor's safety stack is bypassed.
Operational OversightIs our Human-in-the-Loop empowered with a legal mandate to veto, or are they a rubber stamp?Statutory independence: authorised safety officers with protection against operational pressure to override red lines.
Fiduciary LiabilityHave we audited the technical lineage of our high-risk decisions, or are we relying on third-party self-assessments?Conformity evidence: documented, machine-readable proof of data quality, bias mitigation, and robustness.
Geopolitical StrategyDo we have a "cyber emergency" protocol that accounts for sovereign overrides?Multi-vendor verification: secondary, independent AI guardrails to verify high-stakes outputs.

Need Board-Ready Assessment Support?

Run this assessment with June and Charlotte in 15 minutes.

A confidential board risk briefing — no pitch deck, no sales process. An honest answer about your specific exposure.

Book a Board Briefing
06 · Action Plan

What Directors Should Do Next

Audit your AI vendor contracts

Review for "lawful purpose" language that may grant override authority. Your company — not OpenAI — is liable for an illegal data transfer. General Counsel should immediately review Standard Contractual Clauses (SCCs) to assess whether they withstand the reality of the Pentagon's "Master Key."

Request a technical briefing from your CTO/CISO

Ask specifically whether your data isolation is a "soft lock" (software permission) or a "hard lock" (hardware/cryptographic control). These are not the same thing.

Appoint or verify an AI Safety Officer

This person must have genuine independence and legal protection — not report through the CTO chain of command, which creates a direct conflict of interest.

Map your data pipeline

Identify precisely where proprietary data leaves your perimeter and enters a third-party environment. Most boards are surprised by how many exit points exist.

Prepare for August 2026

The EU AI Act becomes fully applicable to high-risk systems. The burden of proof shifts to the board to demonstrate operational evidence of human oversight — not just policy documents.

Members Only

Board Meeting Script: Challenging the “Soft Lock”

Exact questions to raise during your “AI Update,” “Risk Committee,” or “Technology Strategy” agenda items.

  • 5 phases: safety stack, kill switch, HITL, board skills, AI resolution
  • Satisfactory vs red flag answer guide for each question
  • Recommended board resolution language
  • Follow-up frameworks if red flags are found
Access as a Member

Full course access from £3,995

07 · Mitigation

Recommended Mitigation Strategies

The goal is to decouple your company's “Crown Jewels” — intellectual property and private data — from the US military's “Master Key.”

01

Implement Data Residency (Geofencing)

Digital data typically flows to the cheapest or fastest server, often in the US. "Geofencing" forces your data to stay within a specific legal jurisdiction — keeping German data in Frankfurt, for example. This ensures local privacy laws apply and makes it legally harder for the US Pentagon to pull that data under domestic US warrants.

Action

Mandate that high-risk data be processed only in Non-US Regions via Azure OpenAI or AWS Bedrock.

02

Use Zero-Retention Private Environments

By default, AI providers may see or save your data to improve their models or for safety monitoring. A private environment (VPC) creates a digital vault where the AI provider — and by extension, their government partners — cannot access your inputs.

Action

Transition from public chat interfaces to Enterprise Private Instances with Zero Data Retention (ZDR) contractually guaranteed.

03

Adopt "Sovereign AI" for Core IP

If you use a US cloud-based AI, you are subject to US laws like the Defence Production Act. "Sovereign AI" means running AI models on your own physical servers or within a local, domestic cloud in your jurisdiction. If the US government demands access, they have no plug to pull.

Action

For high-value R&D or trade secrets, utilise Open-Weights Models (such as Mistral or Llama) hosted on your own infrastructure.

04

Model Diversification — The "Anthropic Hedge"

Anthropic was blacklisted for refusing to lower their privacy standards. By not being the de facto defence contractor, Anthropic is a less attractive target for foreign intelligence agencies looking to poison military AI supply chains.

Action

Direct the CTO to maintain vendor neutrality so the company can migrate away from any provider that triggers a regulatory red flag.

Members Only

Director's Cheat Sheet: How to Audit the AI Auditors

As August 2026 approaches, audit firms will sell static compliance documents — compliance theatre. This cheat sheet gives directors 6 tests to distinguish genuine preventative governance, including a weighted provider scoring matrix.

  • 6 vetting tests with pass/fail criteria
  • Provider scoring matrix (weighted, 1–5 scale)
  • "Compliance theatre" red flags at a glance
  • 4 next steps for directors
Access as a Member

Full course access from £3,995

References & Sources

  1. 1.OpenAI. "Our agreement with the Department of War." 28 Feb 2026.
  2. 2.TechCrunch. "OpenAI's Sam Altman announces Pentagon deal with technical safeguards." 28 Feb 2026.
  3. 3.TechCrunch. "OpenAI reveals more details about its agreement with the Pentagon." 1 Mar 2026.
  4. 4.Axios. "OpenAI-Pentagon deal faces same safety concerns that plagued Anthropic talks." 1 Mar 2026.
  5. 5.Axios. "Trump moves to blacklist Anthropic's Claude from government work." 27 Feb 2026.
  6. 6.Fortune. "OpenAI sweeps in to snag Pentagon contract." 28 Feb 2026.
  7. 7.CBS News. "Hegseth declares Anthropic a supply chain risk." 28 Feb 2026.
  8. 8.NPR. "OpenAI announces Pentagon deal after Trump bans Anthropic." 27 Feb 2026.
  9. 9.New York Times. "OpenAI Amends AI Deal With the Pentagon." 2 Mar 2026.
  10. 10.NIST AI Risk Management Framework.
  11. 11.EU AI Act — European Commission Digital Strategy.

June Lai (CFA, CPA, CMA, BSc Biochemistry) is Head of AI Governance at AIBoardCourse.com. She advises boards internationally on AI risk management, with qualifications spanning scientific research, finance, and corporate governance.

Next Step

Your Board Needs to Lead on AI — Not Just Respond to It

The AI Board Course gives directors the language, frameworks, and governance tools to govern AI with confidence. Full access to the members-only resources in this article — and every module that follows.