Model AI Legislation Framework
Tier 1 – Foundational Framework | Tier 2 – Technical Basis | Tier 3 – Adoption & Implementation
Tier 1 – Foundational Framework
Version 2.2 | December 2025
A risk-based legislative foundation for artificial intelligence oversight
Developed by AI Safety International
I. Purpose and Legislative Intent
A. Purpose
- Establish a practical, risk-based framework for the evaluation and governance of artificial intelligence systems.
- Provide a standardized failure-analysis method applicable across agencies and jurisdictions.
- Enable oversight without prescribing specific technologies or architectures.
B. Legislative Intent
- This framework is intended to:
- Support innovation while mitigating identifiable harms
- Avoid ideological, ethical, or speculative mandates
- Emphasize measurable risk, documentation, and accountability
- The legislature does not seek to:
- Regulate speech, ideas, or research
- Mandate surveillance of users
- Freeze technological development
II. Scope and Applicability
A. Systems Covered
- Artificial intelligence systems that:
- Engage in direct human interaction
- Influence decision-making, behavior, or emotional state
- Are deployed at scale or within regulated environments
B. Systems Excluded
- Non-interactive computational tools
- Internal research systems not exposed to the public
- Deterministic software without adaptive behavior
C. Tiered Applicability
- Proportional requirements based on:
- Risk classification
- Deployment context
- User exposure level
III. Definitions
A. Artificial Intelligence System
- Functional definition based on behavior and deployment, not model type
B. Failure Mode
- A specific, identifiable way an AI system may produce harm or unintended consequence
C. Risk Assessment
- A structured evaluation of severity, likelihood, and detectability of failure modes
D. Operator / Deployer
- The entity responsible for deployment, configuration, or public availability
IV. Risk-Based Assessment Requirement
A. Mandatory Risk Assessment
- Covered systems must undergo a documented risk analysis prior to deployment
B. Assessment Methodology
- Risk assessment shall:
- Identify plausible failure modes
- Score severity, occurrence, and detectability
- Assign proportional mitigation requirements
C. Accepted Methodologies
- Use of recognized engineering frameworks, including:
- Failure Mode and Effects Analysis (FMEA) or equivalent structured methods
V. Documentation and Accountability
A. Required Documentation
- Risk assessment summary
- Mitigation strategies
- Monitoring and review procedures
B. Record Retention
- Documentation retained for a defined period
- Available for regulatory or judicial review when required
C. Transparency Without Disclosure
- No requirement to disclose proprietary models or training data
- Focus on process and outcomes, not internals
VI. Oversight and Review Mechanisms
A. Designated Oversight Authority
- Assignment to an existing agency or designated body
- No creation of new surveillance structures by default
B. Periodic Review
- Risk assessments updated when:
- System behavior materially changes
- Deployment context changes
- New failure modes are identified
C. Incident Response
- Defined procedures for responding to demonstrated harm
VII. Enforcement and Proportional Remedies
A. Enforcement Principles
- Proportional, risk-based enforcement
- Focus on correction, not punishment
B. Remedies
- Required mitigation
- Temporary deployment limitations
- Documentation correction
C. Safe Harbor
- Good-faith compliance with assessment requirements provides liability mitigation
VIII. Adaptability and Future Updates
A. Framework Evolution
- Allows updates to assessment standards without statutory rewrite
B. Standards Referencing
- Enables incorporation of updated technical standards over time
C. Sunset Review
- Periodic legislative review of framework effectiveness
IX. Non-Preemption and Jurisdictional Coordination
A. Federal / State Compatibility
- Framework designed to coexist with state or sector-specific laws
B. International Alignment
- Compatible with international risk-based approaches without enforcing them
X. Effective Date and Transitional Provisions
A. Phased Implementation
- Grace period for compliance
- Pilot or voluntary adoption options
B. Guidance Period
- Agencies may issue interpretive guidance during rollout
XI. Technical Appendix (Referenced, Not Codified)
A. Risk Assessment Models
- Example AI-FMEA structures
- Severity / Occurrence / Detection scales
B. Illustrative Use Cases
- Conversational AI
- Decision support systems
- Emotional or behavioral influence systems
C. Non-Binding Nature
- Appendix is explanatory, not statutory
Proceed to Tier 2: Technical Basis, which provides the detailed risk-assessment methodology, scoring logic, and operational structures referenced in this framework.
Tier 1 – Foundational Framework | Tier 2 – Technical Basis | Tier 3 – Adoption & Implementation
Printed or downloaded copies may not reflect the most current revision. The authoritative version is maintained at aisafetyinternational.com.
