Back to Cases
Case ReportCompleted2025
AAB-CASE-2025-RV-008

Breakable Machine: A K-12 Classroom Game for Transformative AI Literacy Through Spoofing and eXplainable AI (XAI)

K-12 classroom game resource that teaches critical and transformative AI literacy by having students spoof an image classifier and inspect saliency with XAI, emphasizing model brittleness, bias, and sociotechnical reflection.

This page documents an AI literacy or AI education case for registry purposes. It is descriptive and does not imply AAB endorsement of any specific tool, provider, or intervention.
01

Implementation

University-school collaboration (resource design for K-12 AI education)

02

Learning context

In-school (K-12)

03

AI role

Evaluator

04

Outcome signal

Not specified

Registry Facets

0
Case Type
  • Research Review
Setting
  • K-12
Status
  • Completed
Focus
  • Transformative AI Literacy
  • XAI
  • Adversarial Inquiry

Implementing Organization

1
Organization Type

University-school collaboration (resource design for K-12 AI education)

Location

Finland (designed for broader K-12 classroom use)

Primary Facilitator Role

Researchers and teachers co-developing a classroom AI spoofing/XAI game

Learning Context

2
Setting Type
  • In-school (K-12)
Session Format

Classroom game with collaborative leaderboard and guided reflection

Duration

Classroom session integrated into broader AI project units

Group Size

Whole-class parallel play with individual or small-group participation

Devices

Teacher projector app plus student camera-enabled devices via browser

Constraints
  • Learners need prior exposure to basic image-classification concepts.
  • Activity effectiveness depends on teacher-facilitated reflection beyond gameplay.
  • Hardware/network access and classroom orchestration can affect implementation quality.

Learner Profile

3
Age Range

Grades 4-9 (approximately ages 10-15)

Prior AI Exposure Assumed

Beginner-to-intermediate familiarity with classification concepts

Prior Programming Background Assumed

No coding required for game participation

Educational Intent

4
Primary Learning Goals
  • Understand that high AI confidence does not guarantee correctness.
  • Experience adversarial spoofing to reveal classifier fragility and failure modes.
  • Develop critical reasoning about AI systems as human-made sociotechnical artifacts.
Secondary Learning Goals
  • Interpret feature saliency through XAI heatmaps.
  • Discuss fairness, accountability, and real-world consequences of misclassification.
  • Build learner agency to question and reshape AI-mediated practices.
What This Was Not
  • Not a model-building-only coding lesson.
  • Not a benchmark study of classification accuracy improvements.
  • Not a claim that spoofing itself is the final learning endpoint.

AI Tool Description

5
Tool Type

Browser-based spoofing game with XAI visual explanations and class leaderboard

Languages

Visual interaction and teacher-facilitated classroom discussion

AI Role
  • Evaluator
User Interaction Model
  • Teacher launches challenge labels and class controls through projector interface.
  • Students manipulate appearance/background to induce high-confidence misclassification.
  • Learners inspect CAM heatmaps to infer which visual features drive predictions.
  • Class compares top spoofing strategies and reasoning via shared leaderboard.
Safeguards
  • Local-session data handling with no persistent identifiable storage outside classroom.
  • Privacy-conscious design aligned with GDPR-oriented principles in deployment context.
  • Teacher-mediated reflection to avoid normalizing unsafe or unethical AI misuse.

Activity Design

6
Activity Flow
  • Students join by QR code and receive target classification challenge.
  • Students iterate spoofing attempts while observing confidence feedback.
  • Students switch to XAI heatmap and training-data views for evidence gathering.
  • Classroom discussion links observed failures to robustness, bias, and accountability.
Human Vs AI Responsibilities
  • Humans design challenge conditions, interpret outcomes, and guide ethical inquiry.
  • AI model produces probabilistic labels and saliency outputs.
  • Learners critically evaluate model behavior and propose sociotechnical implications.
Scaffolding Strategies
  • Adversarial play reframes model failure as a pedagogical investigation target.
  • Peer-visible leaderboard supports collaborative hypothesis testing and sensemaking.
  • XAI visualizations make hidden model attention patterns discussable in class.

Observed Challenges

7
Educators Reported
  • Students may over-focus on game performance unless reflection is intentionally scaffolded.
  • Conceptual transfer from spoofing tactics to ethical/systemic reasoning needs facilitation.
  • Classroom logistics (device readiness, pacing, projection visibility) can constrain flow.

Design Adaptations

8
Adaptations
  • Included dual interfaces (teacher control + student interaction) for classroom manageability.
  • Added heatmap and training-data views to move from play to explanation.
  • Used shared leaderboard to externalize successful strategies and foster collective inquiry.
  • Positioned activity as part of broader pedagogical models (e.g., CEDE) instead of standalone use.

Reported Outcomes

9
Engagement
  • Game format promotes high participation through playful challenge and peer comparison.
  • Embodied experimentation supports active inquiry into model behavior.
Learning Signals
  • Learners can observe that semantically irrelevant visual cues can drive predictions.
  • Students encounter and reason about confidence-vs-correctness mismatches.
  • Activity opens pathways to discuss bias, accountability, and societal impact of AI errors.
Educators Reflection

This resource is best used as a bridge from technical curiosity to critical and transformative AI literacy, where failure analysis becomes core content rather than a side effect.

Ethical & Privacy Considerations

10
Privacy
  • Design emphasizes classroom-local data handling and automatic session-end deletion.
  • Learners are prompted to examine who is harmed when AI misclassifies people in real contexts.
  • Critical discussion targets representational harm, dataset assumptions, and governance implications.

Evidence Type

11
Evidence
  • Activity documentation
  • Practitioner observation

Relevance to Research

12
Potential Research Use
  • Provides a concrete model for teaching AI robustness and adversarial behavior in K-12.
  • Expands AI literacy pedagogy from model-building toward critical system interrogation.
  • Offers a practical XAI-mediated entry point for discussing fairness and accountability.
Relevant Research Domains
  • Transformative AI literacy
  • K-12 XAI pedagogy
  • Adversarial thinking in education
  • Data agency and sociotechnical critique

Case Status

13
Case Status
  • Completed

AAB Classification Tags

14
Age

Grades 4-9 (10-15 years)

Setting

Classroom game-based AI inquiry

AI Function

Image classification, saliency explanation, and spoofing analysis

Pedagogy

Adversarial play with collaborative reflection and critical discussion

Risk Level

Medium

Data Sensitivity

Low to Medium (in-session camera inputs and classroom-generated traces)

Registry Metadata

15
Case ID
AAB-CASE-2025-RV-008
Publication Status
Completed
Tags
caseFinland (designed for broader K-12 classroom use)In-school (K-12)