Checking if the AI is Safe for Children
AI Education for Children
Evaluate AI systems used by or around children for safeguarding, developmental, and ethical risk.

Risk Dimensions
Key areas where AI systems may pose risks to children.
Safeguarding
Is the system designed to protect children from harm, exploitation, or inappropriate content?
Developmental appropriateness
Does the system account for cognitive, emotional, and social development stages?
Autonomy and consent
Can children and guardians meaningfully consent to and control the system?
Fairness and inclusion
Does the system treat all children equitably regardless of background or ability?
Transparency
Can the system explain its decisions in age-appropriate language?
Human oversight
Do responsible adults retain meaningful control over the system?
How It Works
Four stages from scoping to actionable recommendations.
Define boundaries
Establish what the AI system should never do. Identify what matters most for child safety.
Envision better design
What would a responsible, child-centred version of this system look like?
Gather perspectives
Consult children directly. What do they experience? Where do they feel safe or unsafe?
Document findings
Compile the audit report with evidence, risk scores, and actionable recommendations.
Ready to improve your product?
Turn audit findings into child-centred design improvements with our research and design service.