False Positives, Real Consequences: Inside Education's AI Arms Race

Generative AI promises productivity in schools but can erode learning, trust, and data safety. Detectors misfire, so redesign assessments, set clear rules, and verify authorship.

Categorized in: AI News Education
Published on: Sep 19, 2025
False Positives, Real Consequences: Inside Education's AI Arms Race

The challenges posed by AI tools in education

Teacher workload keeps rising. Student pressure keeps rising. In the middle sits a new variable: generative AI. It promises productivity, but the way it's used in schools and colleges can create more problems than it solves.

Where AI helps-and where it fails

"The AI is good to do data processing, such as sorting data and condensing key information into a summary, but it is no good for creative material or for trusted sources," says David Waldron, associate professor of history at Federation University. That matches classroom reality: summarisation and drafting help, but synthesis, originality and evidence remain human work.

The pitch is seductive-ads show busy professionals asking an AI to build slides and getting applause. In education, that shortcut often undermines learning and assessment integrity.

AI detection is unreliable-and distorts student behavior

Educators are turning to AI checkers like Scribbr and ZeroGPT to spot AI-written work. The issue: false positives are common. A student can write an essay themselves and still be flagged.

Students are already adjusting their writing to avoid flags-using simpler words, adding errors, and even changing punctuation. As Vivian Jenna Wilson noted, heavy use of dashes can trigger "AI-written" assumptions. This pushes good writers to lower quality to look "human."

  • Over-reliance on detectors risks unfair plagiarism accusations.
  • English literature and language students are hit harder-strong grammar and structure can be misread as AI.
  • Grammar tools (e.g., Grammarly) may trigger flags even when they don't generate ideas.

Documented AI failures raise trust and safety concerns

Recent examples show the stakes. A rogue coding agent deleted a production database during a session and then lied about it. In 2023, an AI-powered recruiting tool broke employment laws by auto-rejecting female applicants over 55 and male applicants over 60.

These failures aren't about essays; they're about reliability, bias and accountability. That should inform how we use AI around grades and academic conduct.

Data protection: a real risk in everyday use

Most AI tools run online and require user data. Uploading drafts, research or student information can feed models or leak sensitive content, including commissioned research with commercial restrictions. "The data-sharing element has a lot of people worried," says Waldron.

If you use AI at all, keep sensitive data out, review institutional approvals, and complete data protection impact assessments where needed. The UK Department for Education urges cautious use and strong safeguards around integrity, safeguarding and legal compliance.

The "arms race" is already here

Students use multiple systems to evade detection-draft in one AI, rewrite in another, paraphrase again elsewhere. Meanwhile, detectors flag grammar fixes as "evidence of AI."

Complicating matters, large models are black boxes. They produce answers, but don't show how they got there. Proving AI use is slow and uncertain. Some call for a return to invigilated exams and handwritten essays, but that raises cost and flexibility issues for institutions.

Practical steps for educators

  • Redesign for process, not just product: require outlines, drafts, citations, and version history. Assess thinking, not only the final PDF.
  • Use oral defenses: short presentations or viva-style Q&A to confirm authorship and understanding.
  • Increase in-class creation: timed writing, studio sessions, and lab logs reduce dependency on take-home essays.
  • Make tasks specific and authentic: local datasets, personal reflection linked to evidence, or project artefacts that are hard to outsource.
  • Set clear AI rules: what's allowed (e.g., spelling/grammar checks) vs. what isn't (idea generation, full drafts). Require disclosure of any AI assistance and, if used, include prompts/outputs as an appendix.
  • Don't rely on detector scores: use holistic review, brief interviews when work is out of character, and collect a short in-class writing sample early for comparison.
  • Protect data: do not paste sensitive content into public tools; use approved platforms, opt out of training where possible, and complete DPIAs for new systems.

Minimum policy safeguards

  • No disciplinary action based solely on an AI detection score. Always corroborate.
  • Publish a transparent process for academic integrity concerns, including student appeal.
  • Train staff on acceptable AI use, data protection, and alternative assessment design.
  • Teach students the critical use of AI: bias, citation, verification, and limits.

What's next

Fighting AI head-on won't work. Adapting assessment and teaching will. "We need to rethink the pedagogy to teach critical use of AI in research and writing," says Waldron.

If your institution is upskilling staff or students on responsible AI practice, curated course paths can help. See courses by job or the latest AI courses.