top of page

AI Use Is Surging in Schools But Where Are the Safeguards?

Updated: Jan 5


By Ben | Taught AI Academy


AI has officially entered the classroom. In 2025, the majority of UK teachers now say they’re using AI tools regularly—whether for planning, differentiation, admin, or communication. And confidence is growing fast: 87% of educators now say they feel either completely or somewhat comfortable using AI at work.


But here’s the concern: the policies, guidance, and safeguarding protections that should sit alongside AI use are either missing—or unknown.


And the data is alarming.


What the Numbers Tell Us

A recent national survey asked over 10,000 teachers whether their school has a policy on staff AI use.


Here’s what they said:

  • Only 15% of primary and 21% of secondary teachers say their school has an AI policy and they know what it says.

  • 65% of primary and 51% of secondary teachers say their school has no AI policy at all.

  • 13% (primary) and 17% (secondary) simply don’t know if a policy exists.

  • Even among those who say their school has a policy, many admit they haven’t read it or can’t remember what it says.


A Growing Safeguarding Risk

This policy vacuum poses a serious safeguarding risk. AI tools are now being used to write parent emails, generate feedback, create lesson content—and yes, sometimes with real pupil data entered into the prompt.

Without clear, enforced guardrails, there’s nothing stopping:

  • Sensitive information being shared with public models

  • Pupil names being entered into unsecured AI systems

  • Staff unknowingly breaching GDPR or safeguarding protocols

  • Conflicting messages about what “safe use” actually looks like


Let’s be blunt: if 1 in 6 teachers doesn’t know whether a policy even exists, the chances of misuse—unintentional or not—are already high.


Teachers Are Not “AI Shy.” They’re Just Undersupported

Interestingly, the survey also revealed that teachers are not scared of AI—they’re open to it. In fact, 87% describe themselves as “completely” or “somewhat” comfortable using AI in their work.


Confidence also rises significantly when staff have read and understood their school’s AI policy:

  • 74% of teachers who had read their school’s AI policy felt “completely comfortable” discussing AI with colleagues.

  • This drops to just 53% among those who don’t know whether a policy exists.

This shows the power of clear guidance: teachers don’t need pushing—they need supporting.


What Schools Should Do Now

This is a moment of both urgency and opportunity. Schools must act—before policy gaps become safeguarding incidents. Here’s what I recommend:


1. Write a Clear, Staff-Facing AI Policy

Cover safeguarding, GDPR, usage boundaries, age-specific rules, and clear do/don’t guidance.

2. Deliver CPD on Safe Prompting & Use

Teachers should know how to use AI tools in a way that’s aligned with school values, inclusive practice, and legal responsibilities.

3. Audit Current AI Usage in Your School

Find out what tools staff are using, how they’re using them, and where the risks (or benefits) lie.

4. Appoint an AI Lead or Working Group

Every school needs a go-to team to review tools, test prompts, support staff, and update policy regularly.


Final Thought: AI Isn’t the Problem. Silence Is.


Teachers are embracing AI. But we’re sending them into this new world without a map.

If schools want to keep children safe, support staff, and use AI meaningfully, they must stop treating AI as a future issue. It’s already here. What’s missing is clear leadership, transparent policy, and proper training.


Because safeguarding isn’t optional. And in the AI age, silence is a risk we can’t afford.



 
 
 

Recent Posts

See All
Your School Improvement Plan Just Got Smarter

By Ben | Taught AI Academy School leaders spend huge amounts of time writing and updating documents like the School Improvement Plan (SIP) , School Development Plan (SDP) , curriculum maps and subject

 
 
 

Comments


bottom of page