Skip to main content

GDS Service Standard Assessment Preparation

The /arckit.service-assessment command helps UK Government teams prepare for GDS Service Standard assessments by analyzing existing ArcKit artifacts as evidence against all 14 points.

What is the Service Standard?

The GDS Service Standard is a set of 14 criteria that UK Government services must meet to achieve alpha, beta, and live assessments. All public-facing services must pass these assessments before launch.

Command: /arckit.service-assessment

Usage

/arckit.service-assessment PHASE=alpha DATE=2025-12-15
Arguments:
  • PHASE (required): alpha, beta, or live - The assessment phase to prepare for
  • DATE (optional): YYYY-MM-DD - Planned assessment date for timeline calculations

Output: ARC-{PROJECT_ID}-SVCASS-v1.0.md

Generates a comprehensive Service Standard assessment preparation report.

The 14-Point Service Standard

The assessment analyzes evidence for all 14 points:

Section 1: Meeting Users’ Needs

  1. Understand users and their needs - Research with diverse users
  2. Solve a whole problem for users - End-to-end user journeys
  3. Provide a joined up experience across all channels - Multi-channel consistency
  4. Make the service simple to use - Usability testing and task completion
  5. Make sure everyone can use the service - WCAG 2.1 AA accessibility

Section 2: Providing a Good Service

  1. Have a multidisciplinary team - Sustainable team with right skills
  2. Use agile ways of working - Iterative delivery and continuous improvement
  3. Iterate and improve frequently - Regular releases based on feedback
  4. Create a secure service which protects users’ privacy - Security and GDPR compliance
  5. Define what success looks like and publish performance data - KPIs and metrics

Section 3: Using the Right Technology

  1. Choose the right tools and technology - Build vs buy analysis
  2. Make new source code open - Open source by default
  3. Use and contribute to open standards, common components and patterns - GOV.UK Design System
  4. Operate a reliable service - Uptime, monitoring, incident response

Evidence Mapping

The command automatically maps ArcKit artifacts to Service Standard points:
Service Standard PointArcKit ArtifactsEvidence Types
1. Understand usersARC-*-STKE-*.md, ARC-*-REQ-*.mdUser research, personas, user stories
2. Solve whole problemARC-*-REQ-*.md, wardley-maps/End-to-end journeys, integration points
3. Joined up experienceARC-*-REQ-*.md, diagrams/Multi-channel requirements, data consistency
4. Simple to useARC-*-REQ-*.md, reviews/ARC-*-HLDR-*.mdUsability NFRs, design review
5. Everyone can useARC-*-REQ-*.md, ARC-*-SECD-*.mdWCAG 2.1 AA requirements, accessibility testing
6. Multidisciplinary teamARC-*-STKE-*.md, ARC-*-PLAN-*.mdRACI matrix, team structure
7. Agile waysARC-*-PLAN-*.mdSprint structure, ceremonies
8. Iterate frequentlyreviews/ARC-*-HLDR-*.md, reviews/ARC-*-DLDR-*.mdDesign iterations, version history
9. Secure and privateARC-*-SECD-*.md, ARC-*-DATA-*.md, ARC-*-DPIA-*.mdSecurity controls, GDPR compliance
10. Success metricsARC-*-REQ-*.md, ARC-*-SOBC-*.mdKPIs, benefits realization
11. Right toolsresearch/, wardley-maps/, ARC-*-TCOP-*.mdTechnology research, build vs buy
12. Open sourcereviews/ARC-*-HLDR-*.md, ARC-*-TCOP-*.mdRepository links, licensing
13. Open standardsARC-*-TCOP-*.md, reviews/ARC-*-HLDR-*.mdGOV.UK Design System, API standards
14. Reliable serviceARC-*-REQ-*.md, reviews/ARC-*-HLDR-*.mdAvailability NFRs, resilience architecture

RAG Ratings

Each point receives a RAG (Red/Amber/Green) rating:
  • 🟒 Green (Ready): All critical evidence found, no significant gaps, ready for assessment
  • 🟑 Amber (Partial): Some evidence but gaps remain, 1-2 weeks to address
  • πŸ”΄ Red (Not Ready): Critical evidence missing, 3+ weeks work required
Overall Readiness:
  • 🟒 Green: 12+ points Green, max 2 Amber, 0 Red
  • 🟑 Amber: 10+ points Green/Amber, max 2 Red
  • πŸ”΄ Red: More than 2 Red points

Report Contents

The assessment preparation report includes:
  1. Executive Summary
    • Overall readiness rating and score (X/14)
    • Critical gaps requiring immediate action
    • Key strengths to showcase
    • Recommended timeline
  2. Detailed Assessment (for each of 14 points)
    • Status: 🟒/🟑/πŸ”΄
    • What this point means and why it matters
    • Evidence required for this phase
    • Evidence found in ArcKit artifacts (with file references)
    • Gap analysis
    • Recommendations (Critical/High/Medium priority)
    • Assessment day guidance
  3. Evidence Inventory
    • Complete traceability: Service Standard Point β†’ ArcKit Artifacts
    • Status and critical gaps table
  4. Assessment Preparation Checklist
    • Critical actions (0-2 weeks)
    • High priority actions (2-4 weeks)
    • Medium priority actions (4+ weeks)
  5. Assessment Day Preparation
    • Timeline and booking guidance
    • Documentation to share with panel (1 week before)
    • Who should attend (core team and phase-specific roles)
    • Show and tell structure (4-hour timeline)
    • Tips for success
    • Materials to have ready
  6. After the Assessment
    • If you pass (Green)
    • If you get Amber (tracking amber evidence process)
    • If you fail (Red) (remediation plan)

Phase-Appropriate Criteria

Alpha Assessment - Focus on demonstrating viability:
  • Lower bar for operational evidence (monitoring, performance data)
  • Higher bar for user research and prototyping
  • Critical: User testing, team composition, technology viability
  • Optional: Full accessibility audit, published performance data
Beta Assessment - Focus on demonstrating production readiness:
  • Higher bar for everything
  • Critical: Working service, security testing, accessibility compliance, performance monitoring
  • All 14 points must be addressed substantively
Live Assessment - Focus on demonstrating continuous improvement:
  • Highest bar, operational excellence expected
  • Critical: Published performance data, user satisfaction, continuous improvement
  • Evidence of service evolution based on user feedback

Example Evidence Requirements

Point 1: Understand Users (Alpha)

βœ… Required:
  • User needs documented from research
  • User groups and personas identified
  • Prototype testing results with real users
  • Evidence of research with diverse user groups
⚠️ Optional:
  • Analytics data (not expected at alpha)

Point 5: Accessibility (Beta)

βœ… Critical:
  • WCAG 2.1 AA audit completed and passed
  • Testing with screen readers, voice control, magnification
  • Testing with disabled users
  • Accessibility statement published
❌ Blocker if missing

Point 10: Performance Data (Live)

βœ… Mandatory:
  • Performance data published on GOV.UK
  • 4 mandatory KPIs: cost per transaction, user satisfaction, completion rate, digital take-up
  • Data updated regularly (at least quarterly)

Integration with Other Commands

The assessment preparation works best when combined with:
  • /arckit.tcop - Technology Code of Practice assessment (Points 11, 13 overlap)
  • /arckit.analyze - Comprehensive governance quality analysis
  • /arckit.traceability - Requirements traceability matrix
  • /arckit.secure - Security assessment (Point 9 evidence)
  • /arckit.requirements - If user stories or NFRs weak
  • /arckit.hld-review - Architecture decisions (Point 11)

Resources

Official GDS Guidance: Phase-Specific Guidance:

Tips for Assessment Day

Do:
  • βœ… Show real work, not polished presentations
  • βœ… Have doers present their work
  • βœ… Be honest about unknowns
  • βœ… Explain problem-solving approach
  • βœ… Demonstrate iteration based on learning
  • βœ… Reference ArcKit artifacts by name
Don’t:
  • ❌ Over-prepare presentations (panel wants artifacts)
  • ❌ Hide problems or pretend everything is perfect
  • ❌ Use jargon or assume panel knows your context
  • ❌ Let senior leaders dominate (panel wants doers)
  • ❌ Argue with panel feedback
  • ❌ Rush through - panel will interrupt with questions

Re-running for Progress Tracking

Re-run the command weekly as you address gaps:
/arckit.service-assessment PHASE=beta
The command will:
  • Detect new evidence added since last run
  • Update RAG ratings based on current state
  • Recalculate overall readiness score
  • Adjust recommendations based on progress
This helps you track progress toward assessment readiness and ensures no critical gaps remain.

Build docs developers (and LLMs) love