Organizations developing or deploying AI systems in the United States must navigate a fragmented regulatory landscape consisting of federal executive orders, voluntary frameworks, state-level AI statutes, and existing federal agency enforcement under consumer protection, employment discrimination, and sector-specific laws. There is no single comprehensive federal AI statute as of March 2026. Instead, compliance requires mapping applicable obligations across three layers: (1) the federal executive order framework, which favors minimal regulation and federal preemption of state laws; (2) state AI statutes -- with California SB 53 imposing frontier model safety requirements with up to $1M per violation, Colorado's AI Act targeting algorithmic discrimination (effective June 30, 2026), Texas TRAIGA with $10K-$200K penalties, and Illinois HB 3773 requiring AI hiring disclosure -- all effective January 1, 2026; and (3) existing enforcement by the FTC, EEOC, and sector regulators who apply decades-old statutes to AI-specific harms. [src1, src2, src3]
As of March 2026, all 50 states, Puerto Rico, the Virgin Islands, and Washington D.C. introduced AI legislation in 2025, with approximately 100 measures adopted or enacted across 38 states. Key state laws now in force or pending include: California SB 53/TFAIA (frontier AI transparency, penalties up to $1M/violation, effective January 1, 2026), California SB 942 (AI content disclosure for systems with 1M+ monthly users, $5,000/day per violation, effective August 2, 2026), Texas TRAIGA ($10,000-$200,000 per violation, effective January 1, 2026), Colorado AI Act ($20,000 per violation, effective June 30, 2026, delayed from February 1, 2026), Illinois HB 3773 (AI employment discrimination, effective January 1, 2026), and New York's RAISE Act ($30M for repeat violations, effective January 1, 2027). NYC Local Law 144, requiring bias audits for automated employment decision tools, has been enforced since July 2023 with penalties of $500-$1,500 per violation per day. The FTC brought five enforcement actions in its September 2024 "Operation AI Comply" sweep targeting deceptive AI claims, though it set aside its Rytr consent order in December 2025. The EEOC issued Title VII guidance in May 2023 establishing that employers are liable for disparate impact from AI hiring tools, even when provided by third-party vendors. [src1, src3, src4, src5, src7, src8]
The US AI regulatory landscape reflects a fundamental tension between innovation promotion and harm prevention that has produced a fragmented patchwork rather than a unified framework. The current administration's December 2025 executive order employs multiple levers to assert federal primacy -- a DOJ litigation task force, FCC preemption proceedings, FTC policy guidance, and $42 billion in broadband funding conditioned on state compliance -- while explicitly preserving state authority over child safety and government procurement. States, meanwhile, continue to fill the perceived federal vacuum: California's SB 53 sets the first US frontier-model safety standard, Colorado targets algorithmic discrimination in high-risk decisions, and Illinois addresses AI in employment. This dual dynamic creates compliance complexity for organizations operating across multiple jurisdictions, particularly since the December 2025 executive order cannot itself preempt state law -- only Congress or the courts can do that. [src1, src2, src3, src7]
START -- Organization needs US AI compliance guidance
|-- What type of AI system?
| |-- Frontier AI model (10^26+ FLOP training)
| | |-- Developer revenue >= $500M?
| | | |-- YES --> California SB 53 full obligations (safety framework, quarterly risk assessments, incident reporting)
| | | +-- NO --> California SB 53 base obligations (transparency report, incident reporting)
| |-- Employment/hiring decision tool
| | |-- Operating in NYC?
| | | |-- YES --> LL144 bias audit required (annual, independent auditor)
| | | +-- NO --> Check state laws (Illinois HB 3773 disclosure) + EEOC Title VII guidance
| | +-- EEOC four-fifths rule applies regardless of location
| |-- Consumer-facing AI product/service
| | |-- 1M+ monthly California users? --> SB 942 disclosure (effective Aug 2026)
| | |-- Making deceptive AI claims? --> FTC Section 5 enforcement risk
| | +-- Operating in California? --> AB 2013 transparency disclosure
| |-- High-risk consequential decision AI
| | |-- Operating in Colorado? --> CAIA applies (June 30, 2026)
| | |-- Operating in Texas? --> TRAIGA applies (Jan 1, 2026)
| | +-- Operating in Utah? --> AI Policy Act applies (May 2024)
| +-- General AI development
| +-- NIST AI RMF voluntary adoption recommended
|-- Is the organization a developer, deployer, or both?
| |-- Developer --> Colorado/Texas/California developer obligations
| |-- Deployer --> Colorado/Texas deployer obligations
| +-- Both --> Full set of obligations applies
+-- Is there an existing AI governance program?
|-- YES --> Audit against NIST AI RMF + applicable state requirements
+-- NO --> Start with NIST AI RMF governance and risk mapping
Organizations that read the December 2025 executive order as immediately invalidating state AI requirements and halt compliance efforts. The EO directs the AG to form a task force and the Commerce Secretary to identify laws by March 2026, but no state law has been preempted as of March 2026. The EO itself expressly states it does not preempt otherwise lawful state AI laws in child safety, data center infrastructure, and government procurement. [src2, src7]
Continue compliance with all applicable state AI laws while monitoring the AI Litigation Task Force actions, FCC preemption proceedings, and any BEAD funding conditions. State laws remain enforceable until a court rules otherwise. Only Congress can formally preempt state law. Treat preemption as a potential future development, not a current reality. [src1, src3]
Organizations that implement the NIST AI Risk Management Framework and assume it satisfies all legal requirements. NIST AI RMF is voluntary and non-binding; it does not substitute for specific statutory obligations under state laws or federal anti-discrimination requirements. [src6]
Adopt NIST AI RMF as a best-practice baseline (it provides safe harbor under Texas TRAIGA), then layer mandatory requirements from applicable state laws, FTC guidance, and EEOC standards on top. Document where voluntary framework compliance meets or exceeds legal obligations. [src6, src3]
Organizations that procure third-party AI hiring tools and assume the vendor handles bias audit compliance for NYC LL144, Illinois HB 3773 disclosure, or disparate impact analysis for EEOC purposes. The EEOC has explicitly stated employers are liable even when AI tools are provided by outside vendors. [src5]
Require vendors to provide bias audit reports and disparate impact analyses, then validate independently. Under LL144, the employer -- not the vendor -- must ensure an independent bias audit exists and publish results. Under Illinois HB 3773, the employer must disclose AI use in hiring to candidates. Under EEOC guidance, the employer bears liability for disparate impact regardless of vendor assurances. [src1, src5, src8]
Organizations that train models at 10^26+ FLOP but assume SB 53 does not apply because they are headquartered outside California. SB 53 applies to frontier developers regardless of location if their models are deployed or accessible in California. [src8]
If the organization trains models at or above the 10^26 FLOP threshold and those models are used by or accessible to California users, SB 53 obligations likely apply. Large frontier developers ($500M+ revenue) face additional requirements including quarterly catastrophic risk reporting. [src8]
Misconception: The US has no AI regulation because Congress has not passed a comprehensive AI law.
Reality: While no single federal AI statute exists, organizations face binding obligations from state laws (California SB 53, Colorado, Texas, Illinois, NYC), existing federal anti-discrimination statutes (Title VII, ADA) as applied to AI by the EEOC, FTC enforcement under Section 5, and sector-specific regulations (HIPAA, FCRA, ECOA). All 50 states introduced AI bills in 2025, with approximately 100 measures adopted across 38 states. [src1, src3, src8]
Misconception: The Biden AI Executive Order 14110 established binding requirements that are now fully repealed.
Reality: EO 14110 primarily directed federal agencies to take actions; its January 2025 revocation removed those agency mandates but did not affect the NIST AI RMF (which predates the EO), state laws, or existing FTC/EEOC enforcement authority. The NIST framework remains fully in effect as a voluntary standard. [src2, src6]
Misconception: Texas TRAIGA and Colorado AI Act have the same requirements.
Reality: They differ significantly. Texas TRAIGA prohibits specific "restricted purposes" (discrimination, rights infringement, deepfake CSAM), provides a 60-day cure period, offers safe harbors for NIST AI RMF compliance, and carries $10K-$200K penalties. Colorado's AI Act focuses on "high-risk" systems making "consequential decisions," requires annual impact assessments, vests exclusive enforcement with the AG (with a 60-day cure notice), penalizes up to $20K per violation, and does not provide a framework-based safe harbor. [src3, src8]
Misconception: California SB 53 only applies to California-based companies.
Reality: SB 53 applies to any "frontier developer" training models at 10^26+ FLOP, regardless of headquarters location, if the model is deployed or accessible in California. This potentially captures all major frontier AI developers. Penalties reach $1M per violation enforced by the California Attorney General. [src8]
Misconception: The December 2025 executive order preempts state AI laws.
Reality: The EO itself cannot preempt state law -- only Congress or the courts can do that. The EO creates mechanisms for future preemption challenges (DOJ task force, FCC proceedings) and economic pressure ($42B BEAD funding conditions), but as of March 2026 no state law has actually been preempted. The EO also explicitly preserves state authority over child safety, data center infrastructure, and government procurement. [src2, src7]
| Rule/Framework | Key Difference | When to Use |
|---|---|---|
| US AI Regulation (this unit) | Fragmented patchwork; no single statute; federal preemption risk; state-by-state obligations | AI deployment in the United States |
| EU AI Act | Risk-tiered comprehensive statute; binding; fines up to EUR 35M or 7% turnover; high-risk obligations delayed to Aug 2027 | AI systems deployed in EU or affecting EU residents |
| NIST AI RMF | Voluntary framework; no legal penalties; four functions (GOVERN, MAP, MEASURE, MANAGE) | As governance foundation in any jurisdiction; safe harbor under Texas TRAIGA |
| California SB 53 (TFAIA) | Frontier AI models only (10^26+ FLOP); safety frameworks; incident reporting; $1M/violation | Developers training large AI models accessible in California |
| Colorado AI Act | High-risk AI focus; deployer impact assessments; $20K/violation; effective June 2026 | AI making consequential decisions in Colorado |
| Texas TRAIGA | Prohibited purposes; 60-day cure; NIST safe harbor; $10K-$200K penalties | AI development or deployment in Texas |
| NYC Local Law 144 | Narrow scope (employment AI only); annual bias audit; NYC jurisdiction only | Automated employment decision tools used in New York City |
Fetch this when a user asks about AI compliance obligations in the United States, US AI executive orders, state AI laws (California SB 53, Colorado AI Act, Texas TRAIGA, Illinois HB 3773), FTC or EEOC enforcement against AI systems, whether AI hiring tools require bias audits, frontier AI model safety requirements, whether federal preemption has eliminated state AI requirements, or how to build an AI governance program for a US-based organization.