UK Government AI Playbook Assessment
Assess UK Government AI Playbook compliance for responsible AI deployment, covering 10 core principles and 6 ethical themes.Command
Arguments
- system (required): AI system name or project identifier
Examples
Purpose
Help UK government organizations assess compliance with the UK Government AI Playbook for responsible AI deployment.Risk Level Determination
HIGH-RISK AI (requires strictest oversight)
Fully automated decisions affecting:- Health and safety
- Fundamental rights
- Access to services
- Legal status
- Employment
- Financial circumstances
MEDIUM-RISK AI (significant impact with human oversight)
- Semi-automated decisions with human review
- Significant resource allocation
LOW-RISK AI (productivity/administrative)
- Recommendation systems with human control
- Administrative automation
The 10 Core Principles
- Understanding AI - Team understands AI limitations and realistic expectations
- Lawful and Ethical Use - DPIA, EqIA, Human Rights assessment completed
- Security - Cyber security assessment including AI-specific threats
- Human Control - Meaningful human oversight and override capability
- Lifecycle Management - Documented lifecycle plan from selection to decommissioning
- Right Tool Selection - Problem clearly defined, alternatives considered
- Collaboration - Cross-government collaboration and knowledge sharing
- Commercial Partnership - Procurement includes AI-specific contract terms
- Skills and Expertise - Team composition includes AI/ML, ethical AI, legal expertise
- Organizational Alignment - AI Governance Board approval and SRO assigned
The 6 Ethical Themes
- Safety, Security, and Robustness - Safety testing, fail-safe mechanisms
- Transparency and Explainability - ATRS published, decision explanations available
- Fairness, Bias, and Discrimination - Bias assessment, fairness metrics across protected characteristics
- Accountability and Responsibility - Clear ownership, audit trail, incident response
- Contestability and Redress - Right to contest AI decisions, appeal mechanism
- Societal Wellbeing and Public Good - Positive societal impact, environmental consideration
Human Oversight Models
- Human-in-the-loop: Review EVERY decision (required for high-risk)
- Human-on-the-loop: Periodic/random review
- Human-in-command: Can override at any time
- Fully automated: AI acts autonomously (HIGH-RISK - justify!)
Output
GeneratesARC-{PROJECT_ID}-AIPB-v{VERSION}.md with:
- Executive summary with overall score (X/160 points, Y%) and Go/No-Go decision
- 10 Principles assessment (each scored 0-10)
- 6 Ethical Themes assessment (each scored 0-10)
- Risk-based decision (HIGH/MEDIUM/LOW risk thresholds)
- Mandatory documentation checklist (ATRS, DPIA, EqIA, etc.)
- Action plan with priorities
- Links to existing ArcKit artifacts
Risk-Based Decision Criteria
- HIGH-RISK: MUST score ≥90%, ALL principles met, human-in-the-loop REQUIRED
- MEDIUM-RISK: SHOULD score ≥75%, critical principles met
- LOW-RISK: SHOULD score ≥60%, basic safeguards in place
Mandatory Documentation
ATRS (Algorithmic Transparency Recording Standard)
- MANDATORY for central government departments
- MANDATORY for arm’s length bodies
- Publish on department website
- Update when system changes significantly
DPIA (Data Protection Impact Assessment)
- MANDATORY for AI processing personal data
- Must be completed BEFORE deployment
- Must be reviewed and updated regularly
EqIA (Equality Impact Assessment)
- MANDATORY to assess impact on protected characteristics
- Must document how discrimination is prevented
Human Rights Assessment
- MANDATORY for decisions affecting rights
- Must consider ECHR (European Convention on Human Rights)
- Document how rights are protected
Prerequisites
MANDATORY (warn if missing):- PRIN (Architecture Principles) - AI governance standards, compliance requirements
- REQ (Requirements) - AI/ML-related FR, NFR (security, fairness), data requirements
- DATA (Data Model) - Training data sources, data quality
- RISK (Risk Register) - AI safety risks, mitigation strategies
Guidance for High-Risk AI
STOP: Do NOT deploy without meeting ALL principles:- Human-in-the-loop MANDATORY (review every decision)
- ATRS publication MANDATORY
- DPIA, EqIA, Human Rights assessments MANDATORY
- Quarterly audits REQUIRED
- AI Governance Board approval REQUIRED
- Senior leadership sign-off REQUIRED
Related Commands
arckit atrs- Generate ATRS record (mandatory for AI systems)arckit dpia- Data Protection Impact Assessmentarckit tcop- Technology Code of Practice (broader tech governance)arckit secure- Security assessment (AI-specific threats)