How ChatGPT Is Quietly Rewriting the Way We Teach Reading in Britain—and What It Means for Every Child

2025-11-26 22:15:22
-

As a member of a UK academic committee responsible for reviewing education innovation and safeguarding academic standards, I have spent the past two years watching the rapid evolution of artificial intelligence with both curiosity and caution. But no development has travelled more quickly from technological curiosity to classroom reality than the use of AI—especially ChatGPT—to generate reading-comprehension questions.

Across the UK, from Year 4 classrooms in Newcastle to GCSE revision groups in London, teachers are increasingly experimenting with AI systems that can, in seconds, produce comprehension activities that once required hours of careful planning. The transformation has been swift, sometimes chaotic, occasionally worrying—and absolutely impossible to ignore.

In this commentary, I want to unpack how this shift happened, what it means for British education, and why the debate now unfolding will shape the literacy of an entire generation. Some will argue that “question generation” is a trivial subfield of AI. In reality, it cuts to the heart of how we teach children to read, think, and understand the world.

And that makes it one of the most important discussions of our time.

51336_9odk_3266.png

The Accidental Revolution: How Teachers Began Using ChatGPT for Reading Tasks

When ChatGPT first launched in late 2022, most educators were concerned that pupils would use it to write essays. Few predicted that teachers would quickly discover a far more practical application: generating the comprehension questions needed to support daily reading practice.

Reading comprehension tasks are the backbone of English literacy instruction. Teachers routinely create:

  • factual recall questions

  • inference questions

  • vocabulary prompts

  • multiple-choice quizzes

  • open-ended analytical tasks

  • short written responses

  • scaffolded sentence starters

For decades, these were written manually. A teacher might spend 30 minutes drafting high-quality questions for a single text—and do this daily.

Suddenly, ChatGPT could produce the same set in fifteen seconds.

The speed was irresistible. It freed teachers from repetitive work. It allowed more tailored content. And it arrived at a moment when schools were already facing overwhelming pressure—from OFSTED requirements, staffing shortages, and the need to support pupils whose literacy suffered during pandemic disruptions.

It was, in other words, the perfect storm for rapid adoption.

Why Teachers Found AI-Generated Questions So Appealing

When asked why they use ChatGPT for reading comprehension, teachers offer three consistent reasons.

1. Time Savings

A Year 6 teacher in Leeds told our committee that AI tools saved her “at least six hours a week”. For a profession in which evenings and weekends are already swallowed by marking, planning, and data recording, saving six hours is nothing short of transformative.

2. Differentiation at Scale

Producing three versions of a comprehension task—for lower-attaining, middle-attaining, and high-attaining pupils—can be exhausting. AI enables customised sets in minutes.

3. Accessibility Adjustments

Teachers can request:

  • simplified texts

  • dyslexia-friendly versions

  • expanded glossaries

  • culturally responsive adaptations

  • translations for EAL learners

  • audio narration support (when combined with text-to-speech)

This creates a level of inclusivity that many schools have long struggled to achieve consistently.

For many educators, these benefits feel like a long-awaited relief.

But There Is a Catch—and It’s Bigger Than Most People Realise

The moment an AI system becomes responsible for shaping what pupils read, it becomes a de-facto co-teacher. And like any teacher, it can get things wrong.

Sometimes spectacularly.

1. Hallucinations

AI models occasionally invent details not present in the original text, leading to inaccurate comprehension questions. If a story mentions a girl walking her dog, the AI may generate a question referring to a cat. For young children still learning to interpret texts, such errors are harmful.

2. Misaligned Difficulty Levels

AI often struggles to match the cognitive difficulty required for SATs, GCSE English Language, or other assessments. Questions may be either too literal or too advanced.

3. Cultural or Contextual Misunderstandings

ChatGPT occasionally produces questions that rely on American idioms, references, or assumptions that do not align with British curriculum expectations.

4. Bias in Text and Question Framing

AI reflects the biases of its training data. That means depictions of characters, settings, or moral choices may inadvertently reinforce stereotypes.

5. Lack of Pedagogical Awareness

Though ChatGPT generates syntactically correct questions, it does not inherently “understand” the learning progression required to build reading competency. It may skip crucial intermediate steps or overemphasise certain skills while ignoring others.

These are not minor issues. They strike at the core of literacy development.

A Quiet Debate Within Academia: Are AI-Generated Questions Even Pedagogically Sound?

Within our academic committee, we have debated for months a deceptively simple question:

Can an AI model produce reading-comprehension tasks that genuinely support learning, rather than simply “look like” they support learning?

Experts are divided.

The Optimists

They argue that:

  • AI improves efficiency

  • teachers retain oversight

  • models can be fine-tuned to UK curricula

  • the benefits outweigh the risks

Some even predict that AI-based question generation could lead to more personalised literacy instruction than previously imaginable.

The Sceptics

They counter that:

  • comprehension questions are scaffolds for thought

  • poor scaffolds produce superficial reading habits

  • AI lacks a theory of mind or narrative understanding

  • literacy outcomes may suffer silently over time

To the sceptics, the fact that AI can produce plausible questions is not evidence that it produces educationally effective ones.

I find myself somewhere in the middle.

The Real Issue: Reading Comprehension Is Not About Questions—It’s About Thinking

If AI merely replaces human-written questions with machine-written ones, nothing significant changes.

But the deeper danger is that schools begin to conflate the appearance of comprehension with the experience of comprehension.

A well-crafted question does more than check understanding; it guides the learner toward cognitive processes:

  • inference

  • prediction

  • character motivation

  • authorial intention

  • structural analysis

  • comparison across texts

  • moral reasoning

AI, by contrast, often produces questions that mimic surface structure without embedding cognitive intention. They may appear correct but fail to cultivate the mental habits that define true literacy.

As one literacy researcher told our committee:

“An AI can generate the shell of a question, but it cannot embed the pedagogy inside it. And pedagogy is where the learning lives.”

This is the heart of the debate.

Parents Are Unaware of the Scale of This Shift

Perhaps the most surprising aspect of this transformation is how little the general public knows about it. Parents across the UK regularly help their children rehearse reading-comprehension activities at home, unaware that many of those questions were generated by an AI system.

If you ask parents whether they want AI-generated homework, many hesitate. They worry about accuracy, fairness, and screen-mediated education.

But they are rarely asked.

This disconnect between classroom practice and parental awareness raises difficult ethical questions. Should schools be required to disclose when AI plays a role in instructional design? Should parents have the right to opt out? Should AI-generated tasks be labelled?

These questions remain unresolved.

The Moral Dimension: Who Controls the Storylines of Our Children?

Literacy education is a form of cultural transmission. Through reading, children internalise:

  • values

  • social norms

  • empathy

  • moral frameworks

  • identity narratives

When AI systems generate reading passages and questions, they indirectly shape these cultural elements.

This raises profound moral considerations:

  • Whose voices are represented in AI-generated content?

  • Whose stories are marginalised or misrepresented?

  • What cultural assumptions are encoded into the prompts?

  • Should corporate AI systems have influence over children’s moral development?

For a nation that values educational autonomy, these questions matter.

Will AI Reduce Teachers to “Editors” Instead of Experts?

Many teachers tell us they appreciate AI-generated materials but fear becoming dependent on them. Several expressed concern that new teachers may lose the ability to design effective tasks from scratch.

This echoes the calculator debate in maths education: efficiency comes at the risk of eroding foundational skills.

Some headteachers have floated the idea of “AI-proof teacher training”, ensuring that pedagogical expertise remains intact even as tools evolve. This may eventually become essential.

The Biggest Opportunity: AI Could Democratising High-Quality Literacy Instruction

For all the concerns, it would be a mistake to overlook the enormous potential. Used well, ChatGPT can:

  • expand access to tailored reading support

  • help struggling readers build confidence

  • reduce teacher workload

  • enable rapid remediation

  • generate culturally aware materials for diverse communities

  • allow pupils to practise independently at home

  • support adults in literacy programmes

  • strengthen consistency across multi-academy trusts

For pupils with dyslexia, ADHD, or language barriers, AI-generated scaffolds could be life-changing.

But only if we use them thoughtfully.

So Where Do We Go From Here? A Framework for Responsible Use

After extensive research, expert interviews, classroom visits, and review of teacher experiences, our committee has begun sketching out a framework for responsible adoption. It rests on six principles.

1. Teacher Oversight Must Remain Central

AI can support, but never replace, professional judgement.

2. Transparency with Parents and Pupils

Schools should disclose when AI-generated material is used, and why.

3. Quality Assurance Standards

Questions must be checked for:

  • factual accuracy

  • alignment with UK curricula

  • cognitive appropriateness

  • cultural sensitivity

  • bias reduction

4. Pedagogical Rigor Over Convenience

Teachers should evaluate not just whether questions are “correct”, but whether they foster meaningful learning.

5. Training and Accreditation

Teachers need structured training on:

  • prompt design

  • bias detection

  • output evaluation

  • pedagogical integration

6. Clear Ethical Boundaries

AI must not be used to subtly shape belief systems, cultural interpretations, or political attitudes in reading materials.

These principles are a starting point, not a conclusion.

A Glimpse Into the Future: What AI Might Soon Be Able to Do

In the next five years, AI systems may be able to:

  • generate comprehension tasks mapped precisely to KS1–KS4 progression

  • analyse pupil responses and diagnose reading difficulties

  • adapt questions in real time

  • create personalised reading journeys

  • annotate texts with cognitive scaffolds

  • identify gaps in reasoning

  • offer teacher-assistive explanations

  • co-design curriculum sequences

Some of this is inspiring. Some is unnerving. All of it demands careful governance.

The Debate We Should Be Having as a Nation

At present, conversations about AI in schools happen in scattered pockets: a staff room here, a multi-academy trust conference there, a worried parent forum online. What we lack is a national conversation about how AI will redefine literacy.

The stakes are enormous.

Reading comprehension is not merely an academic skill. It influences:

  • employment

  • civic participation

  • political literacy

  • emotional development

  • cultural integration

  • social mobility

If AI reshapes how children learn to interpret texts, it may reshape the future of British society itself.

That debate deserves attention—from policymakers, journalists, parents, and the public.

Conclusion: AI Will Transform Reading—but It Must Not Replace Human Thought

ChatGPT has become an invisible collaborator in British classrooms. Sometimes it enhances learning. Sometimes it introduces risk. Most of the time, its influence is unnoticed.

But education is too important to leave to chance.

If we want AI to elevate literacy rather than erode it, we must treat reading-comprehension question generation not as a technical novelty, but as a profound educational intervention that requires oversight, transparency, and scholarly vigilance.

The future of reading in Britain will be shaped not only by algorithms, but by the values we embed into their use.

And we owe it to every child to get it right.