About
AI Federal Framework

AI Federal Framework

Tracking how the White House, Congress, and federal agencies are building a national AI framework - preemption of state law, sectoral regulation through existing agencies, and industry coordination.

10 entries in In-House Counsel Tracker

DOJ Intervenes in xAI Lawsuit to Block Colorado's AI Discrimination Law[1][2][3]

xAI filed suit on April 9, 2026, in U.S. District Court for the District of Colorado to block enforcement of Colorado's SB24-205, a comprehensive AI anti-discrimination law scheduled to take effect June 30, 2026. The statute requires developers and deployers of high-risk AI systems—those used in hiring, lending, and admissions decisions—to conduct impact assessments, make disclosures, and implement risk mitigation measures to prevent algorithmic discrimination. Two weeks later, on April 24, the U.S. Department of Justice intervened with its own complaint, arguing the law violates the Equal Protection Clause by compelling demographic adjustments through disparate-impact liability while simultaneously authorizing discrimination through exemptions for diversity initiatives. The court granted DOJ's intervention and issued a stay suspending enforcement pending resolution.

Sanders and AOC call for federal AI moratorium amid regulatory debate

Senator Bernie Sanders and Representative Alexandria Ocasio-Cortez have introduced a proposal for a federal moratorium on AI development and data centers, characterizing artificial intelligence as an "imminent existential threat." The call for restrictions has crystallized a fundamental policy divide: whether AI requires aggressive regulatory intervention or a risk-based approach that permits innovation while addressing specific harms.

White House pushes federal AI review standards to eliminate "ideological bias"

The Trump administration has established federal review procedures for artificial intelligence systems across government agencies through an executive order titled "Preventing Woke AI in the Federal Government," issued in July 2025 alongside America's AI Action Plan. The order requires federal agencies to implement "Unbiased AI Principles" for large language models in procurement decisions. The Office of Management and Budget must issue implementing guidance within 90 days, after which agencies have an additional 90 days to revise existing contracts and adopt compliance procedures.

Alston & Bird Publishes April 2026 AI Quarterly Review of Key U.S. Laws and Policies

Congress moved on two fronts in late March to shape AI regulation. On March 26, bipartisan lawmakers introduced H.R. 8094, the AI Foundation Model Transparency Act, requiring developers of large language models to disclose training methods, purposes, risks, evaluation protocols, and monitoring practices. The bill imposes no affirmative regulation—only disclosure obligations. One week earlier, the Trump Administration released its National Policy Framework for Artificial Intelligence, a non-binding document recommending Congress adopt unified federal standards across seven areas: child protection, AI infrastructure, intellectual property, free speech, innovation, workforce development, and preemption of state law. The framework followed Senator Marsha Blackburn's March 18 discussion draft of the Trump America AI Act, which would codify President Trump's December 2025 executive order directing federal preemption of state AI laws.

Trump Admin Releases National AI Framework on March 20, 2026

On March 20, 2026, the Trump administration released the "National Policy Framework for Artificial Intelligence: Legislative Recommendations," a detailed statutory blueprint that would establish uniform federal AI policy and preempt most state regulations. The Framework, mandated by an December 2025 executive order, proposes that Congress delegate AI development oversight to existing sector-specific agencies rather than create a new federal regulator. It would allow states limited authority only in narrow areas: child safety, fraud prevention, zoning, and government procurement. The administration has tasked the Department of Justice with challenging state AI laws through a dedicated task force, while the Department of Commerce will evaluate state regulations deemed "onerous," and the Federal Trade Commission will enforce preemption policies on deceptive practices.

White House Releases 2026 National AI Policy Framework on March 20

On March 20, 2026, the White House released the National Policy Framework for Artificial Intelligence, proposing federal legislation to preempt state laws that impose "undue burdens" on AI deployment. The framework aims to establish uniform national standards for AI governance across sectors, particularly healthcare, where the technology is rapidly expanding into clinical decision support, diagnostics, and administrative workflows. The initiative follows a December 2025 Executive Order directing the administration to develop coordinated federal policy. Implementation would distribute oversight among existing agencies—the FDA, CMS, HHS, OCR, FTC, and DOJ—rather than creating a new regulatory body. The Department of Commerce would evaluate conflicting state laws.

What Your AI Knows About You

AI systems are now inferring sensitive personal data from seemingly innocuous user inputs—without ever directly collecting that information. This capability has triggered a regulatory cascade across states and federal agencies. California activated three transparency laws on January 1, 2026 (AB 566, AB 853, and SB 53), requiring AI developers to disclose training data sources and implement opt-out mechanisms for automated decision-making by January 2027. Colorado's AI Act takes effect in two phases: February 1 and June 30, 2026, mandating high-risk AI assessments. The EU's AI Act reaches full implementation in August 2026. Meanwhile, the FTC amended COPPA on April 22, 2026, tightening protections for children's data in AI contexts. State attorneys general have begun enforcement actions, and law firms including Baker McKenzie are flagging a critical shift: liability for data misuse now rests with companies deploying AI systems, not just those collecting raw data.

Q1 2026 AI Agents Spark IP Debates in Software Development

In the first quarter of 2026, autonomous AI workflow agents including Openclaw demonstrated the ability to generate production-ready software directly from user specifications. The capability triggered immediate debate over intellectual property ownership, developer liability, and the legal framework governing self-generating code.

FCA Sticks to Existing Rules for AI Oversight in Finance

The UK Financial Conduct Authority has reaffirmed its decision to regulate artificial intelligence in financial services through existing principles-based rules rather than new AI-specific legislation. The FCA is applying its current framework—including the Consumer Duty, Senior Managers and Certification Regime, systems and controls requirements, and operational resilience standards—to firms' design, deployment, and oversight of AI systems. The Prudential Regulation Authority and Bank of England have adopted the same approach, rejecting prescriptive AI rules in favor of technology-agnostic scrutiny of firms' processes.

Anthropic's Mythos AI Preview Gains US Gov't Momentum Despite Risks

On April 20, 2026, Anthropic's Mythos Preview—a frontier AI model—continued operating across U.S. government agencies including the NSA and Department of War despite DoW flagging Anthropic as a supply chain risk. The model's continued deployment underscores its perceived indispensability to federal operations, even as security concerns mount.

LawSnap Briefing Updated May 6, 2026

State of play.

Where things stand.

Latest developments.

Active questions and open splits.

  • Whether the Colorado intervention becomes the federal template. DOJ's Equal Protection theory — that disparate-impact liability compels demographic adjustments — is novel and untested at the appellate level. If it holds, it forecloses most state algorithmic bias statutes; if it fails, state enforcement authority survives (→ DOJ Intervenes in xAI Lawsuit to Block Colorado's AI Discrimination Law[1][2][3]).
  • Whether voluntary pre-release review hardens into mandatory obligation. Every major lab is now in the CAISI program voluntarily; an EO mandating industry-wide review is under active consideration. The line between voluntary and compelled — and what triggers liability for non-participation — is unresolved (→ Google, Microsoft and xAI Agree to Share Early AI Models With U.S.).
  • Preemption scope: development vs. use. The Framework's proposed distinction — federal authority over AI development, state authority over AI use — has not been tested under the major questions doctrine. Whether courts accept that line, and how it maps onto state anti-discrimination and disclosure statutes, is the central constitutional question (→ Trump Admin Releases National AI Framework on March 20, 2026).
  • Federal procurement "ideological neutrality" standard. The "Unbiased AI Principles" requirement creates a procurement compliance obligation whose content OMB has not yet defined. Vendors face the risk of contract revision requirements against an undefined standard (→ White House pushes federal AI review standards to eliminate "ideological bias").
  • IP ownership for AI-generated outputs. No settled rule exists on whether code, analysis, or other outputs generated by autonomous AI agents vest in the user, the developer, or remain unowned. Litigation is anticipated; the National AI Framework addresses the question without resolving it (→ Q1 2026 AI Agents Spark IP Debates in Software Development).
  • Open-source vs. proprietary in national security context. The USCC report showing Chinese open-source models leading on HuggingFace has sharpened the policy debate; the administration has not resolved whether to embrace open models or treat them as a security risk — and the answer directly affects export controls and procurement .
  • EU AI Act August 2026 binding effect vs. US framework divergence. Multinational clients face a compliance fork: the EU's risk-based prescriptive regime binds in August while the US framework remains non-binding legislative recommendations. Whether the two converge on specific standards — or require parallel compliance architectures — is unresolved (→ Alston & Bird Publishes April 2026 AI Quarterly Review of Key U.S. Laws and Policies).

What to watch.

  • Colorado's May 13 deadline for successor legislation — whether the amended statute addresses DOJ's Equal Protection theory or triggers a second intervention.
  • Whether the White House issues an EO mandating industry-wide pre-release AI review, and what enforcement mechanism it carries.
  • OMB publication of the "Unbiased AI Principles" implementing guidance and what it requires of existing federal AI contracts.
  • Whether DOJ's preemption task force files additional interventions against California, New York, or other state AI statutes.
  • EU AI Act binding effect in August 2026 — whether US developers treat it as a compliance floor that implicitly shapes domestic practice.
  • Commerce Department release of its evaluation of "onerous" state AI regulations — the document that was due March 11 and remains unreleased.

mail Subscribe to AI Federal Framework email updates

Primary sources. No fluff. Straight to your inbox.

Also on LawSnap